001 // Generated by the protocol buffer compiler. DO NOT EDIT! 002 // source: QJournalProtocol.proto 003 004 package org.apache.hadoop.hdfs.qjournal.protocol; 005 006 public final class QJournalProtocolProtos { 007 private QJournalProtocolProtos() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public interface JournalIdProtoOrBuilder 012 extends com.google.protobuf.MessageOrBuilder { 013 014 // required string identifier = 1; 015 /** 016 * <code>required string identifier = 1;</code> 017 */ 018 boolean hasIdentifier(); 019 /** 020 * <code>required string identifier = 1;</code> 021 */ 022 java.lang.String getIdentifier(); 023 /** 024 * <code>required string identifier = 1;</code> 025 */ 026 com.google.protobuf.ByteString 027 getIdentifierBytes(); 028 } 029 /** 030 * Protobuf type {@code hadoop.hdfs.JournalIdProto} 031 */ 032 public static final class JournalIdProto extends 033 com.google.protobuf.GeneratedMessage 034 implements JournalIdProtoOrBuilder { 035 // Use JournalIdProto.newBuilder() to construct. 036 private JournalIdProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 037 super(builder); 038 this.unknownFields = builder.getUnknownFields(); 039 } 040 private JournalIdProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 041 042 private static final JournalIdProto defaultInstance; 043 public static JournalIdProto getDefaultInstance() { 044 return defaultInstance; 045 } 046 047 public JournalIdProto getDefaultInstanceForType() { 048 return defaultInstance; 049 } 050 051 private final com.google.protobuf.UnknownFieldSet unknownFields; 052 @java.lang.Override 053 public final com.google.protobuf.UnknownFieldSet 054 getUnknownFields() { 055 return this.unknownFields; 056 } 057 private JournalIdProto( 058 com.google.protobuf.CodedInputStream input, 059 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 060 throws com.google.protobuf.InvalidProtocolBufferException { 061 initFields(); 062 int mutable_bitField0_ = 0; 063 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 064 com.google.protobuf.UnknownFieldSet.newBuilder(); 065 try { 066 boolean done = false; 067 while (!done) { 068 int tag = input.readTag(); 069 switch (tag) { 070 case 0: 071 done = true; 072 break; 073 default: { 074 if (!parseUnknownField(input, unknownFields, 075 extensionRegistry, tag)) { 076 done = true; 077 } 078 break; 079 } 080 case 10: { 081 bitField0_ |= 0x00000001; 082 identifier_ = input.readBytes(); 083 break; 084 } 085 } 086 } 087 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 088 throw e.setUnfinishedMessage(this); 089 } catch (java.io.IOException e) { 090 throw new com.google.protobuf.InvalidProtocolBufferException( 091 e.getMessage()).setUnfinishedMessage(this); 092 } finally { 093 this.unknownFields = unknownFields.build(); 094 makeExtensionsImmutable(); 095 } 096 } 097 public static final com.google.protobuf.Descriptors.Descriptor 098 getDescriptor() { 099 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor; 100 } 101 102 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 103 internalGetFieldAccessorTable() { 104 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable 105 .ensureFieldAccessorsInitialized( 106 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class); 107 } 108 109 public static com.google.protobuf.Parser<JournalIdProto> PARSER = 110 new com.google.protobuf.AbstractParser<JournalIdProto>() { 111 public JournalIdProto parsePartialFrom( 112 com.google.protobuf.CodedInputStream input, 113 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 114 throws com.google.protobuf.InvalidProtocolBufferException { 115 return new JournalIdProto(input, extensionRegistry); 116 } 117 }; 118 119 @java.lang.Override 120 public com.google.protobuf.Parser<JournalIdProto> getParserForType() { 121 return PARSER; 122 } 123 124 private int bitField0_; 125 // required string identifier = 1; 126 public static final int IDENTIFIER_FIELD_NUMBER = 1; 127 private java.lang.Object identifier_; 128 /** 129 * <code>required string identifier = 1;</code> 130 */ 131 public boolean hasIdentifier() { 132 return ((bitField0_ & 0x00000001) == 0x00000001); 133 } 134 /** 135 * <code>required string identifier = 1;</code> 136 */ 137 public java.lang.String getIdentifier() { 138 java.lang.Object ref = identifier_; 139 if (ref instanceof java.lang.String) { 140 return (java.lang.String) ref; 141 } else { 142 com.google.protobuf.ByteString bs = 143 (com.google.protobuf.ByteString) ref; 144 java.lang.String s = bs.toStringUtf8(); 145 if (bs.isValidUtf8()) { 146 identifier_ = s; 147 } 148 return s; 149 } 150 } 151 /** 152 * <code>required string identifier = 1;</code> 153 */ 154 public com.google.protobuf.ByteString 155 getIdentifierBytes() { 156 java.lang.Object ref = identifier_; 157 if (ref instanceof java.lang.String) { 158 com.google.protobuf.ByteString b = 159 com.google.protobuf.ByteString.copyFromUtf8( 160 (java.lang.String) ref); 161 identifier_ = b; 162 return b; 163 } else { 164 return (com.google.protobuf.ByteString) ref; 165 } 166 } 167 168 private void initFields() { 169 identifier_ = ""; 170 } 171 private byte memoizedIsInitialized = -1; 172 public final boolean isInitialized() { 173 byte isInitialized = memoizedIsInitialized; 174 if (isInitialized != -1) return isInitialized == 1; 175 176 if (!hasIdentifier()) { 177 memoizedIsInitialized = 0; 178 return false; 179 } 180 memoizedIsInitialized = 1; 181 return true; 182 } 183 184 public void writeTo(com.google.protobuf.CodedOutputStream output) 185 throws java.io.IOException { 186 getSerializedSize(); 187 if (((bitField0_ & 0x00000001) == 0x00000001)) { 188 output.writeBytes(1, getIdentifierBytes()); 189 } 190 getUnknownFields().writeTo(output); 191 } 192 193 private int memoizedSerializedSize = -1; 194 public int getSerializedSize() { 195 int size = memoizedSerializedSize; 196 if (size != -1) return size; 197 198 size = 0; 199 if (((bitField0_ & 0x00000001) == 0x00000001)) { 200 size += com.google.protobuf.CodedOutputStream 201 .computeBytesSize(1, getIdentifierBytes()); 202 } 203 size += getUnknownFields().getSerializedSize(); 204 memoizedSerializedSize = size; 205 return size; 206 } 207 208 private static final long serialVersionUID = 0L; 209 @java.lang.Override 210 protected java.lang.Object writeReplace() 211 throws java.io.ObjectStreamException { 212 return super.writeReplace(); 213 } 214 215 @java.lang.Override 216 public boolean equals(final java.lang.Object obj) { 217 if (obj == this) { 218 return true; 219 } 220 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)) { 221 return super.equals(obj); 222 } 223 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) obj; 224 225 boolean result = true; 226 result = result && (hasIdentifier() == other.hasIdentifier()); 227 if (hasIdentifier()) { 228 result = result && getIdentifier() 229 .equals(other.getIdentifier()); 230 } 231 result = result && 232 getUnknownFields().equals(other.getUnknownFields()); 233 return result; 234 } 235 236 private int memoizedHashCode = 0; 237 @java.lang.Override 238 public int hashCode() { 239 if (memoizedHashCode != 0) { 240 return memoizedHashCode; 241 } 242 int hash = 41; 243 hash = (19 * hash) + getDescriptorForType().hashCode(); 244 if (hasIdentifier()) { 245 hash = (37 * hash) + IDENTIFIER_FIELD_NUMBER; 246 hash = (53 * hash) + getIdentifier().hashCode(); 247 } 248 hash = (29 * hash) + getUnknownFields().hashCode(); 249 memoizedHashCode = hash; 250 return hash; 251 } 252 253 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 254 com.google.protobuf.ByteString data) 255 throws com.google.protobuf.InvalidProtocolBufferException { 256 return PARSER.parseFrom(data); 257 } 258 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 259 com.google.protobuf.ByteString data, 260 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 261 throws com.google.protobuf.InvalidProtocolBufferException { 262 return PARSER.parseFrom(data, extensionRegistry); 263 } 264 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(byte[] data) 265 throws com.google.protobuf.InvalidProtocolBufferException { 266 return PARSER.parseFrom(data); 267 } 268 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 269 byte[] data, 270 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 271 throws com.google.protobuf.InvalidProtocolBufferException { 272 return PARSER.parseFrom(data, extensionRegistry); 273 } 274 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(java.io.InputStream input) 275 throws java.io.IOException { 276 return PARSER.parseFrom(input); 277 } 278 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 279 java.io.InputStream input, 280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 281 throws java.io.IOException { 282 return PARSER.parseFrom(input, extensionRegistry); 283 } 284 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(java.io.InputStream input) 285 throws java.io.IOException { 286 return PARSER.parseDelimitedFrom(input); 287 } 288 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom( 289 java.io.InputStream input, 290 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 291 throws java.io.IOException { 292 return PARSER.parseDelimitedFrom(input, extensionRegistry); 293 } 294 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 295 com.google.protobuf.CodedInputStream input) 296 throws java.io.IOException { 297 return PARSER.parseFrom(input); 298 } 299 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom( 300 com.google.protobuf.CodedInputStream input, 301 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 302 throws java.io.IOException { 303 return PARSER.parseFrom(input, extensionRegistry); 304 } 305 306 public static Builder newBuilder() { return Builder.create(); } 307 public Builder newBuilderForType() { return newBuilder(); } 308 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto prototype) { 309 return newBuilder().mergeFrom(prototype); 310 } 311 public Builder toBuilder() { return newBuilder(this); } 312 313 @java.lang.Override 314 protected Builder newBuilderForType( 315 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 316 Builder builder = new Builder(parent); 317 return builder; 318 } 319 /** 320 * Protobuf type {@code hadoop.hdfs.JournalIdProto} 321 */ 322 public static final class Builder extends 323 com.google.protobuf.GeneratedMessage.Builder<Builder> 324 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder { 325 public static final com.google.protobuf.Descriptors.Descriptor 326 getDescriptor() { 327 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor; 328 } 329 330 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 331 internalGetFieldAccessorTable() { 332 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable 333 .ensureFieldAccessorsInitialized( 334 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class); 335 } 336 337 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder() 338 private Builder() { 339 maybeForceBuilderInitialization(); 340 } 341 342 private Builder( 343 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 344 super(parent); 345 maybeForceBuilderInitialization(); 346 } 347 private void maybeForceBuilderInitialization() { 348 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 349 } 350 } 351 private static Builder create() { 352 return new Builder(); 353 } 354 355 public Builder clear() { 356 super.clear(); 357 identifier_ = ""; 358 bitField0_ = (bitField0_ & ~0x00000001); 359 return this; 360 } 361 362 public Builder clone() { 363 return create().mergeFrom(buildPartial()); 364 } 365 366 public com.google.protobuf.Descriptors.Descriptor 367 getDescriptorForType() { 368 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor; 369 } 370 371 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getDefaultInstanceForType() { 372 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 373 } 374 375 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto build() { 376 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = buildPartial(); 377 if (!result.isInitialized()) { 378 throw newUninitializedMessageException(result); 379 } 380 return result; 381 } 382 383 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto buildPartial() { 384 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto(this); 385 int from_bitField0_ = bitField0_; 386 int to_bitField0_ = 0; 387 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 388 to_bitField0_ |= 0x00000001; 389 } 390 result.identifier_ = identifier_; 391 result.bitField0_ = to_bitField0_; 392 onBuilt(); 393 return result; 394 } 395 396 public Builder mergeFrom(com.google.protobuf.Message other) { 397 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) { 398 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)other); 399 } else { 400 super.mergeFrom(other); 401 return this; 402 } 403 } 404 405 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other) { 406 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) return this; 407 if (other.hasIdentifier()) { 408 bitField0_ |= 0x00000001; 409 identifier_ = other.identifier_; 410 onChanged(); 411 } 412 this.mergeUnknownFields(other.getUnknownFields()); 413 return this; 414 } 415 416 public final boolean isInitialized() { 417 if (!hasIdentifier()) { 418 419 return false; 420 } 421 return true; 422 } 423 424 public Builder mergeFrom( 425 com.google.protobuf.CodedInputStream input, 426 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 427 throws java.io.IOException { 428 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parsedMessage = null; 429 try { 430 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 431 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 432 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) e.getUnfinishedMessage(); 433 throw e; 434 } finally { 435 if (parsedMessage != null) { 436 mergeFrom(parsedMessage); 437 } 438 } 439 return this; 440 } 441 private int bitField0_; 442 443 // required string identifier = 1; 444 private java.lang.Object identifier_ = ""; 445 /** 446 * <code>required string identifier = 1;</code> 447 */ 448 public boolean hasIdentifier() { 449 return ((bitField0_ & 0x00000001) == 0x00000001); 450 } 451 /** 452 * <code>required string identifier = 1;</code> 453 */ 454 public java.lang.String getIdentifier() { 455 java.lang.Object ref = identifier_; 456 if (!(ref instanceof java.lang.String)) { 457 java.lang.String s = ((com.google.protobuf.ByteString) ref) 458 .toStringUtf8(); 459 identifier_ = s; 460 return s; 461 } else { 462 return (java.lang.String) ref; 463 } 464 } 465 /** 466 * <code>required string identifier = 1;</code> 467 */ 468 public com.google.protobuf.ByteString 469 getIdentifierBytes() { 470 java.lang.Object ref = identifier_; 471 if (ref instanceof String) { 472 com.google.protobuf.ByteString b = 473 com.google.protobuf.ByteString.copyFromUtf8( 474 (java.lang.String) ref); 475 identifier_ = b; 476 return b; 477 } else { 478 return (com.google.protobuf.ByteString) ref; 479 } 480 } 481 /** 482 * <code>required string identifier = 1;</code> 483 */ 484 public Builder setIdentifier( 485 java.lang.String value) { 486 if (value == null) { 487 throw new NullPointerException(); 488 } 489 bitField0_ |= 0x00000001; 490 identifier_ = value; 491 onChanged(); 492 return this; 493 } 494 /** 495 * <code>required string identifier = 1;</code> 496 */ 497 public Builder clearIdentifier() { 498 bitField0_ = (bitField0_ & ~0x00000001); 499 identifier_ = getDefaultInstance().getIdentifier(); 500 onChanged(); 501 return this; 502 } 503 /** 504 * <code>required string identifier = 1;</code> 505 */ 506 public Builder setIdentifierBytes( 507 com.google.protobuf.ByteString value) { 508 if (value == null) { 509 throw new NullPointerException(); 510 } 511 bitField0_ |= 0x00000001; 512 identifier_ = value; 513 onChanged(); 514 return this; 515 } 516 517 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalIdProto) 518 } 519 520 static { 521 defaultInstance = new JournalIdProto(true); 522 defaultInstance.initFields(); 523 } 524 525 // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalIdProto) 526 } 527 528 public interface RequestInfoProtoOrBuilder 529 extends com.google.protobuf.MessageOrBuilder { 530 531 // required .hadoop.hdfs.JournalIdProto journalId = 1; 532 /** 533 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 534 */ 535 boolean hasJournalId(); 536 /** 537 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 538 */ 539 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId(); 540 /** 541 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 542 */ 543 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder(); 544 545 // required uint64 epoch = 2; 546 /** 547 * <code>required uint64 epoch = 2;</code> 548 */ 549 boolean hasEpoch(); 550 /** 551 * <code>required uint64 epoch = 2;</code> 552 */ 553 long getEpoch(); 554 555 // required uint64 ipcSerialNumber = 3; 556 /** 557 * <code>required uint64 ipcSerialNumber = 3;</code> 558 */ 559 boolean hasIpcSerialNumber(); 560 /** 561 * <code>required uint64 ipcSerialNumber = 3;</code> 562 */ 563 long getIpcSerialNumber(); 564 565 // optional uint64 committedTxId = 4; 566 /** 567 * <code>optional uint64 committedTxId = 4;</code> 568 * 569 * <pre> 570 * Whenever a writer makes a request, it informs 571 * the node of the latest committed txid. This may 572 * be higher than the transaction data included in the 573 * request itself, eg in the case that the node has 574 * fallen behind. 575 * </pre> 576 */ 577 boolean hasCommittedTxId(); 578 /** 579 * <code>optional uint64 committedTxId = 4;</code> 580 * 581 * <pre> 582 * Whenever a writer makes a request, it informs 583 * the node of the latest committed txid. This may 584 * be higher than the transaction data included in the 585 * request itself, eg in the case that the node has 586 * fallen behind. 587 * </pre> 588 */ 589 long getCommittedTxId(); 590 } 591 /** 592 * Protobuf type {@code hadoop.hdfs.RequestInfoProto} 593 */ 594 public static final class RequestInfoProto extends 595 com.google.protobuf.GeneratedMessage 596 implements RequestInfoProtoOrBuilder { 597 // Use RequestInfoProto.newBuilder() to construct. 598 private RequestInfoProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 599 super(builder); 600 this.unknownFields = builder.getUnknownFields(); 601 } 602 private RequestInfoProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 603 604 private static final RequestInfoProto defaultInstance; 605 public static RequestInfoProto getDefaultInstance() { 606 return defaultInstance; 607 } 608 609 public RequestInfoProto getDefaultInstanceForType() { 610 return defaultInstance; 611 } 612 613 private final com.google.protobuf.UnknownFieldSet unknownFields; 614 @java.lang.Override 615 public final com.google.protobuf.UnknownFieldSet 616 getUnknownFields() { 617 return this.unknownFields; 618 } 619 private RequestInfoProto( 620 com.google.protobuf.CodedInputStream input, 621 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 622 throws com.google.protobuf.InvalidProtocolBufferException { 623 initFields(); 624 int mutable_bitField0_ = 0; 625 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 626 com.google.protobuf.UnknownFieldSet.newBuilder(); 627 try { 628 boolean done = false; 629 while (!done) { 630 int tag = input.readTag(); 631 switch (tag) { 632 case 0: 633 done = true; 634 break; 635 default: { 636 if (!parseUnknownField(input, unknownFields, 637 extensionRegistry, tag)) { 638 done = true; 639 } 640 break; 641 } 642 case 10: { 643 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 644 if (((bitField0_ & 0x00000001) == 0x00000001)) { 645 subBuilder = journalId_.toBuilder(); 646 } 647 journalId_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 648 if (subBuilder != null) { 649 subBuilder.mergeFrom(journalId_); 650 journalId_ = subBuilder.buildPartial(); 651 } 652 bitField0_ |= 0x00000001; 653 break; 654 } 655 case 16: { 656 bitField0_ |= 0x00000002; 657 epoch_ = input.readUInt64(); 658 break; 659 } 660 case 24: { 661 bitField0_ |= 0x00000004; 662 ipcSerialNumber_ = input.readUInt64(); 663 break; 664 } 665 case 32: { 666 bitField0_ |= 0x00000008; 667 committedTxId_ = input.readUInt64(); 668 break; 669 } 670 } 671 } 672 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 673 throw e.setUnfinishedMessage(this); 674 } catch (java.io.IOException e) { 675 throw new com.google.protobuf.InvalidProtocolBufferException( 676 e.getMessage()).setUnfinishedMessage(this); 677 } finally { 678 this.unknownFields = unknownFields.build(); 679 makeExtensionsImmutable(); 680 } 681 } 682 public static final com.google.protobuf.Descriptors.Descriptor 683 getDescriptor() { 684 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor; 685 } 686 687 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 688 internalGetFieldAccessorTable() { 689 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable 690 .ensureFieldAccessorsInitialized( 691 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class); 692 } 693 694 public static com.google.protobuf.Parser<RequestInfoProto> PARSER = 695 new com.google.protobuf.AbstractParser<RequestInfoProto>() { 696 public RequestInfoProto parsePartialFrom( 697 com.google.protobuf.CodedInputStream input, 698 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 699 throws com.google.protobuf.InvalidProtocolBufferException { 700 return new RequestInfoProto(input, extensionRegistry); 701 } 702 }; 703 704 @java.lang.Override 705 public com.google.protobuf.Parser<RequestInfoProto> getParserForType() { 706 return PARSER; 707 } 708 709 private int bitField0_; 710 // required .hadoop.hdfs.JournalIdProto journalId = 1; 711 public static final int JOURNALID_FIELD_NUMBER = 1; 712 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_; 713 /** 714 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 715 */ 716 public boolean hasJournalId() { 717 return ((bitField0_ & 0x00000001) == 0x00000001); 718 } 719 /** 720 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 721 */ 722 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() { 723 return journalId_; 724 } 725 /** 726 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 727 */ 728 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() { 729 return journalId_; 730 } 731 732 // required uint64 epoch = 2; 733 public static final int EPOCH_FIELD_NUMBER = 2; 734 private long epoch_; 735 /** 736 * <code>required uint64 epoch = 2;</code> 737 */ 738 public boolean hasEpoch() { 739 return ((bitField0_ & 0x00000002) == 0x00000002); 740 } 741 /** 742 * <code>required uint64 epoch = 2;</code> 743 */ 744 public long getEpoch() { 745 return epoch_; 746 } 747 748 // required uint64 ipcSerialNumber = 3; 749 public static final int IPCSERIALNUMBER_FIELD_NUMBER = 3; 750 private long ipcSerialNumber_; 751 /** 752 * <code>required uint64 ipcSerialNumber = 3;</code> 753 */ 754 public boolean hasIpcSerialNumber() { 755 return ((bitField0_ & 0x00000004) == 0x00000004); 756 } 757 /** 758 * <code>required uint64 ipcSerialNumber = 3;</code> 759 */ 760 public long getIpcSerialNumber() { 761 return ipcSerialNumber_; 762 } 763 764 // optional uint64 committedTxId = 4; 765 public static final int COMMITTEDTXID_FIELD_NUMBER = 4; 766 private long committedTxId_; 767 /** 768 * <code>optional uint64 committedTxId = 4;</code> 769 * 770 * <pre> 771 * Whenever a writer makes a request, it informs 772 * the node of the latest committed txid. This may 773 * be higher than the transaction data included in the 774 * request itself, eg in the case that the node has 775 * fallen behind. 776 * </pre> 777 */ 778 public boolean hasCommittedTxId() { 779 return ((bitField0_ & 0x00000008) == 0x00000008); 780 } 781 /** 782 * <code>optional uint64 committedTxId = 4;</code> 783 * 784 * <pre> 785 * Whenever a writer makes a request, it informs 786 * the node of the latest committed txid. This may 787 * be higher than the transaction data included in the 788 * request itself, eg in the case that the node has 789 * fallen behind. 790 * </pre> 791 */ 792 public long getCommittedTxId() { 793 return committedTxId_; 794 } 795 796 private void initFields() { 797 journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 798 epoch_ = 0L; 799 ipcSerialNumber_ = 0L; 800 committedTxId_ = 0L; 801 } 802 private byte memoizedIsInitialized = -1; 803 public final boolean isInitialized() { 804 byte isInitialized = memoizedIsInitialized; 805 if (isInitialized != -1) return isInitialized == 1; 806 807 if (!hasJournalId()) { 808 memoizedIsInitialized = 0; 809 return false; 810 } 811 if (!hasEpoch()) { 812 memoizedIsInitialized = 0; 813 return false; 814 } 815 if (!hasIpcSerialNumber()) { 816 memoizedIsInitialized = 0; 817 return false; 818 } 819 if (!getJournalId().isInitialized()) { 820 memoizedIsInitialized = 0; 821 return false; 822 } 823 memoizedIsInitialized = 1; 824 return true; 825 } 826 827 public void writeTo(com.google.protobuf.CodedOutputStream output) 828 throws java.io.IOException { 829 getSerializedSize(); 830 if (((bitField0_ & 0x00000001) == 0x00000001)) { 831 output.writeMessage(1, journalId_); 832 } 833 if (((bitField0_ & 0x00000002) == 0x00000002)) { 834 output.writeUInt64(2, epoch_); 835 } 836 if (((bitField0_ & 0x00000004) == 0x00000004)) { 837 output.writeUInt64(3, ipcSerialNumber_); 838 } 839 if (((bitField0_ & 0x00000008) == 0x00000008)) { 840 output.writeUInt64(4, committedTxId_); 841 } 842 getUnknownFields().writeTo(output); 843 } 844 845 private int memoizedSerializedSize = -1; 846 public int getSerializedSize() { 847 int size = memoizedSerializedSize; 848 if (size != -1) return size; 849 850 size = 0; 851 if (((bitField0_ & 0x00000001) == 0x00000001)) { 852 size += com.google.protobuf.CodedOutputStream 853 .computeMessageSize(1, journalId_); 854 } 855 if (((bitField0_ & 0x00000002) == 0x00000002)) { 856 size += com.google.protobuf.CodedOutputStream 857 .computeUInt64Size(2, epoch_); 858 } 859 if (((bitField0_ & 0x00000004) == 0x00000004)) { 860 size += com.google.protobuf.CodedOutputStream 861 .computeUInt64Size(3, ipcSerialNumber_); 862 } 863 if (((bitField0_ & 0x00000008) == 0x00000008)) { 864 size += com.google.protobuf.CodedOutputStream 865 .computeUInt64Size(4, committedTxId_); 866 } 867 size += getUnknownFields().getSerializedSize(); 868 memoizedSerializedSize = size; 869 return size; 870 } 871 872 private static final long serialVersionUID = 0L; 873 @java.lang.Override 874 protected java.lang.Object writeReplace() 875 throws java.io.ObjectStreamException { 876 return super.writeReplace(); 877 } 878 879 @java.lang.Override 880 public boolean equals(final java.lang.Object obj) { 881 if (obj == this) { 882 return true; 883 } 884 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)) { 885 return super.equals(obj); 886 } 887 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) obj; 888 889 boolean result = true; 890 result = result && (hasJournalId() == other.hasJournalId()); 891 if (hasJournalId()) { 892 result = result && getJournalId() 893 .equals(other.getJournalId()); 894 } 895 result = result && (hasEpoch() == other.hasEpoch()); 896 if (hasEpoch()) { 897 result = result && (getEpoch() 898 == other.getEpoch()); 899 } 900 result = result && (hasIpcSerialNumber() == other.hasIpcSerialNumber()); 901 if (hasIpcSerialNumber()) { 902 result = result && (getIpcSerialNumber() 903 == other.getIpcSerialNumber()); 904 } 905 result = result && (hasCommittedTxId() == other.hasCommittedTxId()); 906 if (hasCommittedTxId()) { 907 result = result && (getCommittedTxId() 908 == other.getCommittedTxId()); 909 } 910 result = result && 911 getUnknownFields().equals(other.getUnknownFields()); 912 return result; 913 } 914 915 private int memoizedHashCode = 0; 916 @java.lang.Override 917 public int hashCode() { 918 if (memoizedHashCode != 0) { 919 return memoizedHashCode; 920 } 921 int hash = 41; 922 hash = (19 * hash) + getDescriptorForType().hashCode(); 923 if (hasJournalId()) { 924 hash = (37 * hash) + JOURNALID_FIELD_NUMBER; 925 hash = (53 * hash) + getJournalId().hashCode(); 926 } 927 if (hasEpoch()) { 928 hash = (37 * hash) + EPOCH_FIELD_NUMBER; 929 hash = (53 * hash) + hashLong(getEpoch()); 930 } 931 if (hasIpcSerialNumber()) { 932 hash = (37 * hash) + IPCSERIALNUMBER_FIELD_NUMBER; 933 hash = (53 * hash) + hashLong(getIpcSerialNumber()); 934 } 935 if (hasCommittedTxId()) { 936 hash = (37 * hash) + COMMITTEDTXID_FIELD_NUMBER; 937 hash = (53 * hash) + hashLong(getCommittedTxId()); 938 } 939 hash = (29 * hash) + getUnknownFields().hashCode(); 940 memoizedHashCode = hash; 941 return hash; 942 } 943 944 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 945 com.google.protobuf.ByteString data) 946 throws com.google.protobuf.InvalidProtocolBufferException { 947 return PARSER.parseFrom(data); 948 } 949 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 950 com.google.protobuf.ByteString data, 951 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 952 throws com.google.protobuf.InvalidProtocolBufferException { 953 return PARSER.parseFrom(data, extensionRegistry); 954 } 955 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(byte[] data) 956 throws com.google.protobuf.InvalidProtocolBufferException { 957 return PARSER.parseFrom(data); 958 } 959 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 960 byte[] data, 961 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 962 throws com.google.protobuf.InvalidProtocolBufferException { 963 return PARSER.parseFrom(data, extensionRegistry); 964 } 965 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(java.io.InputStream input) 966 throws java.io.IOException { 967 return PARSER.parseFrom(input); 968 } 969 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 970 java.io.InputStream input, 971 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 972 throws java.io.IOException { 973 return PARSER.parseFrom(input, extensionRegistry); 974 } 975 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(java.io.InputStream input) 976 throws java.io.IOException { 977 return PARSER.parseDelimitedFrom(input); 978 } 979 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom( 980 java.io.InputStream input, 981 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 982 throws java.io.IOException { 983 return PARSER.parseDelimitedFrom(input, extensionRegistry); 984 } 985 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 986 com.google.protobuf.CodedInputStream input) 987 throws java.io.IOException { 988 return PARSER.parseFrom(input); 989 } 990 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom( 991 com.google.protobuf.CodedInputStream input, 992 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 993 throws java.io.IOException { 994 return PARSER.parseFrom(input, extensionRegistry); 995 } 996 997 public static Builder newBuilder() { return Builder.create(); } 998 public Builder newBuilderForType() { return newBuilder(); } 999 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto prototype) { 1000 return newBuilder().mergeFrom(prototype); 1001 } 1002 public Builder toBuilder() { return newBuilder(this); } 1003 1004 @java.lang.Override 1005 protected Builder newBuilderForType( 1006 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1007 Builder builder = new Builder(parent); 1008 return builder; 1009 } 1010 /** 1011 * Protobuf type {@code hadoop.hdfs.RequestInfoProto} 1012 */ 1013 public static final class Builder extends 1014 com.google.protobuf.GeneratedMessage.Builder<Builder> 1015 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder { 1016 public static final com.google.protobuf.Descriptors.Descriptor 1017 getDescriptor() { 1018 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor; 1019 } 1020 1021 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1022 internalGetFieldAccessorTable() { 1023 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable 1024 .ensureFieldAccessorsInitialized( 1025 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class); 1026 } 1027 1028 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder() 1029 private Builder() { 1030 maybeForceBuilderInitialization(); 1031 } 1032 1033 private Builder( 1034 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1035 super(parent); 1036 maybeForceBuilderInitialization(); 1037 } 1038 private void maybeForceBuilderInitialization() { 1039 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1040 getJournalIdFieldBuilder(); 1041 } 1042 } 1043 private static Builder create() { 1044 return new Builder(); 1045 } 1046 1047 public Builder clear() { 1048 super.clear(); 1049 if (journalIdBuilder_ == null) { 1050 journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 1051 } else { 1052 journalIdBuilder_.clear(); 1053 } 1054 bitField0_ = (bitField0_ & ~0x00000001); 1055 epoch_ = 0L; 1056 bitField0_ = (bitField0_ & ~0x00000002); 1057 ipcSerialNumber_ = 0L; 1058 bitField0_ = (bitField0_ & ~0x00000004); 1059 committedTxId_ = 0L; 1060 bitField0_ = (bitField0_ & ~0x00000008); 1061 return this; 1062 } 1063 1064 public Builder clone() { 1065 return create().mergeFrom(buildPartial()); 1066 } 1067 1068 public com.google.protobuf.Descriptors.Descriptor 1069 getDescriptorForType() { 1070 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor; 1071 } 1072 1073 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getDefaultInstanceForType() { 1074 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 1075 } 1076 1077 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto build() { 1078 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = buildPartial(); 1079 if (!result.isInitialized()) { 1080 throw newUninitializedMessageException(result); 1081 } 1082 return result; 1083 } 1084 1085 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto buildPartial() { 1086 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto(this); 1087 int from_bitField0_ = bitField0_; 1088 int to_bitField0_ = 0; 1089 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1090 to_bitField0_ |= 0x00000001; 1091 } 1092 if (journalIdBuilder_ == null) { 1093 result.journalId_ = journalId_; 1094 } else { 1095 result.journalId_ = journalIdBuilder_.build(); 1096 } 1097 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 1098 to_bitField0_ |= 0x00000002; 1099 } 1100 result.epoch_ = epoch_; 1101 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 1102 to_bitField0_ |= 0x00000004; 1103 } 1104 result.ipcSerialNumber_ = ipcSerialNumber_; 1105 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 1106 to_bitField0_ |= 0x00000008; 1107 } 1108 result.committedTxId_ = committedTxId_; 1109 result.bitField0_ = to_bitField0_; 1110 onBuilt(); 1111 return result; 1112 } 1113 1114 public Builder mergeFrom(com.google.protobuf.Message other) { 1115 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) { 1116 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)other); 1117 } else { 1118 super.mergeFrom(other); 1119 return this; 1120 } 1121 } 1122 1123 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other) { 1124 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) return this; 1125 if (other.hasJournalId()) { 1126 mergeJournalId(other.getJournalId()); 1127 } 1128 if (other.hasEpoch()) { 1129 setEpoch(other.getEpoch()); 1130 } 1131 if (other.hasIpcSerialNumber()) { 1132 setIpcSerialNumber(other.getIpcSerialNumber()); 1133 } 1134 if (other.hasCommittedTxId()) { 1135 setCommittedTxId(other.getCommittedTxId()); 1136 } 1137 this.mergeUnknownFields(other.getUnknownFields()); 1138 return this; 1139 } 1140 1141 public final boolean isInitialized() { 1142 if (!hasJournalId()) { 1143 1144 return false; 1145 } 1146 if (!hasEpoch()) { 1147 1148 return false; 1149 } 1150 if (!hasIpcSerialNumber()) { 1151 1152 return false; 1153 } 1154 if (!getJournalId().isInitialized()) { 1155 1156 return false; 1157 } 1158 return true; 1159 } 1160 1161 public Builder mergeFrom( 1162 com.google.protobuf.CodedInputStream input, 1163 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1164 throws java.io.IOException { 1165 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parsedMessage = null; 1166 try { 1167 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1168 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1169 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) e.getUnfinishedMessage(); 1170 throw e; 1171 } finally { 1172 if (parsedMessage != null) { 1173 mergeFrom(parsedMessage); 1174 } 1175 } 1176 return this; 1177 } 1178 private int bitField0_; 1179 1180 // required .hadoop.hdfs.JournalIdProto journalId = 1; 1181 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 1182 private com.google.protobuf.SingleFieldBuilder< 1183 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> journalIdBuilder_; 1184 /** 1185 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1186 */ 1187 public boolean hasJournalId() { 1188 return ((bitField0_ & 0x00000001) == 0x00000001); 1189 } 1190 /** 1191 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1192 */ 1193 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() { 1194 if (journalIdBuilder_ == null) { 1195 return journalId_; 1196 } else { 1197 return journalIdBuilder_.getMessage(); 1198 } 1199 } 1200 /** 1201 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1202 */ 1203 public Builder setJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 1204 if (journalIdBuilder_ == null) { 1205 if (value == null) { 1206 throw new NullPointerException(); 1207 } 1208 journalId_ = value; 1209 onChanged(); 1210 } else { 1211 journalIdBuilder_.setMessage(value); 1212 } 1213 bitField0_ |= 0x00000001; 1214 return this; 1215 } 1216 /** 1217 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1218 */ 1219 public Builder setJournalId( 1220 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 1221 if (journalIdBuilder_ == null) { 1222 journalId_ = builderForValue.build(); 1223 onChanged(); 1224 } else { 1225 journalIdBuilder_.setMessage(builderForValue.build()); 1226 } 1227 bitField0_ |= 0x00000001; 1228 return this; 1229 } 1230 /** 1231 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1232 */ 1233 public Builder mergeJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 1234 if (journalIdBuilder_ == null) { 1235 if (((bitField0_ & 0x00000001) == 0x00000001) && 1236 journalId_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 1237 journalId_ = 1238 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(journalId_).mergeFrom(value).buildPartial(); 1239 } else { 1240 journalId_ = value; 1241 } 1242 onChanged(); 1243 } else { 1244 journalIdBuilder_.mergeFrom(value); 1245 } 1246 bitField0_ |= 0x00000001; 1247 return this; 1248 } 1249 /** 1250 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1251 */ 1252 public Builder clearJournalId() { 1253 if (journalIdBuilder_ == null) { 1254 journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 1255 onChanged(); 1256 } else { 1257 journalIdBuilder_.clear(); 1258 } 1259 bitField0_ = (bitField0_ & ~0x00000001); 1260 return this; 1261 } 1262 /** 1263 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1264 */ 1265 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJournalIdBuilder() { 1266 bitField0_ |= 0x00000001; 1267 onChanged(); 1268 return getJournalIdFieldBuilder().getBuilder(); 1269 } 1270 /** 1271 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1272 */ 1273 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() { 1274 if (journalIdBuilder_ != null) { 1275 return journalIdBuilder_.getMessageOrBuilder(); 1276 } else { 1277 return journalId_; 1278 } 1279 } 1280 /** 1281 * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code> 1282 */ 1283 private com.google.protobuf.SingleFieldBuilder< 1284 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 1285 getJournalIdFieldBuilder() { 1286 if (journalIdBuilder_ == null) { 1287 journalIdBuilder_ = new com.google.protobuf.SingleFieldBuilder< 1288 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 1289 journalId_, 1290 getParentForChildren(), 1291 isClean()); 1292 journalId_ = null; 1293 } 1294 return journalIdBuilder_; 1295 } 1296 1297 // required uint64 epoch = 2; 1298 private long epoch_ ; 1299 /** 1300 * <code>required uint64 epoch = 2;</code> 1301 */ 1302 public boolean hasEpoch() { 1303 return ((bitField0_ & 0x00000002) == 0x00000002); 1304 } 1305 /** 1306 * <code>required uint64 epoch = 2;</code> 1307 */ 1308 public long getEpoch() { 1309 return epoch_; 1310 } 1311 /** 1312 * <code>required uint64 epoch = 2;</code> 1313 */ 1314 public Builder setEpoch(long value) { 1315 bitField0_ |= 0x00000002; 1316 epoch_ = value; 1317 onChanged(); 1318 return this; 1319 } 1320 /** 1321 * <code>required uint64 epoch = 2;</code> 1322 */ 1323 public Builder clearEpoch() { 1324 bitField0_ = (bitField0_ & ~0x00000002); 1325 epoch_ = 0L; 1326 onChanged(); 1327 return this; 1328 } 1329 1330 // required uint64 ipcSerialNumber = 3; 1331 private long ipcSerialNumber_ ; 1332 /** 1333 * <code>required uint64 ipcSerialNumber = 3;</code> 1334 */ 1335 public boolean hasIpcSerialNumber() { 1336 return ((bitField0_ & 0x00000004) == 0x00000004); 1337 } 1338 /** 1339 * <code>required uint64 ipcSerialNumber = 3;</code> 1340 */ 1341 public long getIpcSerialNumber() { 1342 return ipcSerialNumber_; 1343 } 1344 /** 1345 * <code>required uint64 ipcSerialNumber = 3;</code> 1346 */ 1347 public Builder setIpcSerialNumber(long value) { 1348 bitField0_ |= 0x00000004; 1349 ipcSerialNumber_ = value; 1350 onChanged(); 1351 return this; 1352 } 1353 /** 1354 * <code>required uint64 ipcSerialNumber = 3;</code> 1355 */ 1356 public Builder clearIpcSerialNumber() { 1357 bitField0_ = (bitField0_ & ~0x00000004); 1358 ipcSerialNumber_ = 0L; 1359 onChanged(); 1360 return this; 1361 } 1362 1363 // optional uint64 committedTxId = 4; 1364 private long committedTxId_ ; 1365 /** 1366 * <code>optional uint64 committedTxId = 4;</code> 1367 * 1368 * <pre> 1369 * Whenever a writer makes a request, it informs 1370 * the node of the latest committed txid. This may 1371 * be higher than the transaction data included in the 1372 * request itself, eg in the case that the node has 1373 * fallen behind. 1374 * </pre> 1375 */ 1376 public boolean hasCommittedTxId() { 1377 return ((bitField0_ & 0x00000008) == 0x00000008); 1378 } 1379 /** 1380 * <code>optional uint64 committedTxId = 4;</code> 1381 * 1382 * <pre> 1383 * Whenever a writer makes a request, it informs 1384 * the node of the latest committed txid. This may 1385 * be higher than the transaction data included in the 1386 * request itself, eg in the case that the node has 1387 * fallen behind. 1388 * </pre> 1389 */ 1390 public long getCommittedTxId() { 1391 return committedTxId_; 1392 } 1393 /** 1394 * <code>optional uint64 committedTxId = 4;</code> 1395 * 1396 * <pre> 1397 * Whenever a writer makes a request, it informs 1398 * the node of the latest committed txid. This may 1399 * be higher than the transaction data included in the 1400 * request itself, eg in the case that the node has 1401 * fallen behind. 1402 * </pre> 1403 */ 1404 public Builder setCommittedTxId(long value) { 1405 bitField0_ |= 0x00000008; 1406 committedTxId_ = value; 1407 onChanged(); 1408 return this; 1409 } 1410 /** 1411 * <code>optional uint64 committedTxId = 4;</code> 1412 * 1413 * <pre> 1414 * Whenever a writer makes a request, it informs 1415 * the node of the latest committed txid. This may 1416 * be higher than the transaction data included in the 1417 * request itself, eg in the case that the node has 1418 * fallen behind. 1419 * </pre> 1420 */ 1421 public Builder clearCommittedTxId() { 1422 bitField0_ = (bitField0_ & ~0x00000008); 1423 committedTxId_ = 0L; 1424 onChanged(); 1425 return this; 1426 } 1427 1428 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.RequestInfoProto) 1429 } 1430 1431 static { 1432 defaultInstance = new RequestInfoProto(true); 1433 defaultInstance.initFields(); 1434 } 1435 1436 // @@protoc_insertion_point(class_scope:hadoop.hdfs.RequestInfoProto) 1437 } 1438 1439 public interface SegmentStateProtoOrBuilder 1440 extends com.google.protobuf.MessageOrBuilder { 1441 1442 // required uint64 startTxId = 1; 1443 /** 1444 * <code>required uint64 startTxId = 1;</code> 1445 */ 1446 boolean hasStartTxId(); 1447 /** 1448 * <code>required uint64 startTxId = 1;</code> 1449 */ 1450 long getStartTxId(); 1451 1452 // required uint64 endTxId = 2; 1453 /** 1454 * <code>required uint64 endTxId = 2;</code> 1455 */ 1456 boolean hasEndTxId(); 1457 /** 1458 * <code>required uint64 endTxId = 2;</code> 1459 */ 1460 long getEndTxId(); 1461 1462 // required bool isInProgress = 3; 1463 /** 1464 * <code>required bool isInProgress = 3;</code> 1465 */ 1466 boolean hasIsInProgress(); 1467 /** 1468 * <code>required bool isInProgress = 3;</code> 1469 */ 1470 boolean getIsInProgress(); 1471 } 1472 /** 1473 * Protobuf type {@code hadoop.hdfs.SegmentStateProto} 1474 */ 1475 public static final class SegmentStateProto extends 1476 com.google.protobuf.GeneratedMessage 1477 implements SegmentStateProtoOrBuilder { 1478 // Use SegmentStateProto.newBuilder() to construct. 1479 private SegmentStateProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1480 super(builder); 1481 this.unknownFields = builder.getUnknownFields(); 1482 } 1483 private SegmentStateProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1484 1485 private static final SegmentStateProto defaultInstance; 1486 public static SegmentStateProto getDefaultInstance() { 1487 return defaultInstance; 1488 } 1489 1490 public SegmentStateProto getDefaultInstanceForType() { 1491 return defaultInstance; 1492 } 1493 1494 private final com.google.protobuf.UnknownFieldSet unknownFields; 1495 @java.lang.Override 1496 public final com.google.protobuf.UnknownFieldSet 1497 getUnknownFields() { 1498 return this.unknownFields; 1499 } 1500 private SegmentStateProto( 1501 com.google.protobuf.CodedInputStream input, 1502 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1503 throws com.google.protobuf.InvalidProtocolBufferException { 1504 initFields(); 1505 int mutable_bitField0_ = 0; 1506 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1507 com.google.protobuf.UnknownFieldSet.newBuilder(); 1508 try { 1509 boolean done = false; 1510 while (!done) { 1511 int tag = input.readTag(); 1512 switch (tag) { 1513 case 0: 1514 done = true; 1515 break; 1516 default: { 1517 if (!parseUnknownField(input, unknownFields, 1518 extensionRegistry, tag)) { 1519 done = true; 1520 } 1521 break; 1522 } 1523 case 8: { 1524 bitField0_ |= 0x00000001; 1525 startTxId_ = input.readUInt64(); 1526 break; 1527 } 1528 case 16: { 1529 bitField0_ |= 0x00000002; 1530 endTxId_ = input.readUInt64(); 1531 break; 1532 } 1533 case 24: { 1534 bitField0_ |= 0x00000004; 1535 isInProgress_ = input.readBool(); 1536 break; 1537 } 1538 } 1539 } 1540 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1541 throw e.setUnfinishedMessage(this); 1542 } catch (java.io.IOException e) { 1543 throw new com.google.protobuf.InvalidProtocolBufferException( 1544 e.getMessage()).setUnfinishedMessage(this); 1545 } finally { 1546 this.unknownFields = unknownFields.build(); 1547 makeExtensionsImmutable(); 1548 } 1549 } 1550 public static final com.google.protobuf.Descriptors.Descriptor 1551 getDescriptor() { 1552 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor; 1553 } 1554 1555 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1556 internalGetFieldAccessorTable() { 1557 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable 1558 .ensureFieldAccessorsInitialized( 1559 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class); 1560 } 1561 1562 public static com.google.protobuf.Parser<SegmentStateProto> PARSER = 1563 new com.google.protobuf.AbstractParser<SegmentStateProto>() { 1564 public SegmentStateProto parsePartialFrom( 1565 com.google.protobuf.CodedInputStream input, 1566 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1567 throws com.google.protobuf.InvalidProtocolBufferException { 1568 return new SegmentStateProto(input, extensionRegistry); 1569 } 1570 }; 1571 1572 @java.lang.Override 1573 public com.google.protobuf.Parser<SegmentStateProto> getParserForType() { 1574 return PARSER; 1575 } 1576 1577 private int bitField0_; 1578 // required uint64 startTxId = 1; 1579 public static final int STARTTXID_FIELD_NUMBER = 1; 1580 private long startTxId_; 1581 /** 1582 * <code>required uint64 startTxId = 1;</code> 1583 */ 1584 public boolean hasStartTxId() { 1585 return ((bitField0_ & 0x00000001) == 0x00000001); 1586 } 1587 /** 1588 * <code>required uint64 startTxId = 1;</code> 1589 */ 1590 public long getStartTxId() { 1591 return startTxId_; 1592 } 1593 1594 // required uint64 endTxId = 2; 1595 public static final int ENDTXID_FIELD_NUMBER = 2; 1596 private long endTxId_; 1597 /** 1598 * <code>required uint64 endTxId = 2;</code> 1599 */ 1600 public boolean hasEndTxId() { 1601 return ((bitField0_ & 0x00000002) == 0x00000002); 1602 } 1603 /** 1604 * <code>required uint64 endTxId = 2;</code> 1605 */ 1606 public long getEndTxId() { 1607 return endTxId_; 1608 } 1609 1610 // required bool isInProgress = 3; 1611 public static final int ISINPROGRESS_FIELD_NUMBER = 3; 1612 private boolean isInProgress_; 1613 /** 1614 * <code>required bool isInProgress = 3;</code> 1615 */ 1616 public boolean hasIsInProgress() { 1617 return ((bitField0_ & 0x00000004) == 0x00000004); 1618 } 1619 /** 1620 * <code>required bool isInProgress = 3;</code> 1621 */ 1622 public boolean getIsInProgress() { 1623 return isInProgress_; 1624 } 1625 1626 private void initFields() { 1627 startTxId_ = 0L; 1628 endTxId_ = 0L; 1629 isInProgress_ = false; 1630 } 1631 private byte memoizedIsInitialized = -1; 1632 public final boolean isInitialized() { 1633 byte isInitialized = memoizedIsInitialized; 1634 if (isInitialized != -1) return isInitialized == 1; 1635 1636 if (!hasStartTxId()) { 1637 memoizedIsInitialized = 0; 1638 return false; 1639 } 1640 if (!hasEndTxId()) { 1641 memoizedIsInitialized = 0; 1642 return false; 1643 } 1644 if (!hasIsInProgress()) { 1645 memoizedIsInitialized = 0; 1646 return false; 1647 } 1648 memoizedIsInitialized = 1; 1649 return true; 1650 } 1651 1652 public void writeTo(com.google.protobuf.CodedOutputStream output) 1653 throws java.io.IOException { 1654 getSerializedSize(); 1655 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1656 output.writeUInt64(1, startTxId_); 1657 } 1658 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1659 output.writeUInt64(2, endTxId_); 1660 } 1661 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1662 output.writeBool(3, isInProgress_); 1663 } 1664 getUnknownFields().writeTo(output); 1665 } 1666 1667 private int memoizedSerializedSize = -1; 1668 public int getSerializedSize() { 1669 int size = memoizedSerializedSize; 1670 if (size != -1) return size; 1671 1672 size = 0; 1673 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1674 size += com.google.protobuf.CodedOutputStream 1675 .computeUInt64Size(1, startTxId_); 1676 } 1677 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1678 size += com.google.protobuf.CodedOutputStream 1679 .computeUInt64Size(2, endTxId_); 1680 } 1681 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1682 size += com.google.protobuf.CodedOutputStream 1683 .computeBoolSize(3, isInProgress_); 1684 } 1685 size += getUnknownFields().getSerializedSize(); 1686 memoizedSerializedSize = size; 1687 return size; 1688 } 1689 1690 private static final long serialVersionUID = 0L; 1691 @java.lang.Override 1692 protected java.lang.Object writeReplace() 1693 throws java.io.ObjectStreamException { 1694 return super.writeReplace(); 1695 } 1696 1697 @java.lang.Override 1698 public boolean equals(final java.lang.Object obj) { 1699 if (obj == this) { 1700 return true; 1701 } 1702 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)) { 1703 return super.equals(obj); 1704 } 1705 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) obj; 1706 1707 boolean result = true; 1708 result = result && (hasStartTxId() == other.hasStartTxId()); 1709 if (hasStartTxId()) { 1710 result = result && (getStartTxId() 1711 == other.getStartTxId()); 1712 } 1713 result = result && (hasEndTxId() == other.hasEndTxId()); 1714 if (hasEndTxId()) { 1715 result = result && (getEndTxId() 1716 == other.getEndTxId()); 1717 } 1718 result = result && (hasIsInProgress() == other.hasIsInProgress()); 1719 if (hasIsInProgress()) { 1720 result = result && (getIsInProgress() 1721 == other.getIsInProgress()); 1722 } 1723 result = result && 1724 getUnknownFields().equals(other.getUnknownFields()); 1725 return result; 1726 } 1727 1728 private int memoizedHashCode = 0; 1729 @java.lang.Override 1730 public int hashCode() { 1731 if (memoizedHashCode != 0) { 1732 return memoizedHashCode; 1733 } 1734 int hash = 41; 1735 hash = (19 * hash) + getDescriptorForType().hashCode(); 1736 if (hasStartTxId()) { 1737 hash = (37 * hash) + STARTTXID_FIELD_NUMBER; 1738 hash = (53 * hash) + hashLong(getStartTxId()); 1739 } 1740 if (hasEndTxId()) { 1741 hash = (37 * hash) + ENDTXID_FIELD_NUMBER; 1742 hash = (53 * hash) + hashLong(getEndTxId()); 1743 } 1744 if (hasIsInProgress()) { 1745 hash = (37 * hash) + ISINPROGRESS_FIELD_NUMBER; 1746 hash = (53 * hash) + hashBoolean(getIsInProgress()); 1747 } 1748 hash = (29 * hash) + getUnknownFields().hashCode(); 1749 memoizedHashCode = hash; 1750 return hash; 1751 } 1752 1753 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1754 com.google.protobuf.ByteString data) 1755 throws com.google.protobuf.InvalidProtocolBufferException { 1756 return PARSER.parseFrom(data); 1757 } 1758 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1759 com.google.protobuf.ByteString data, 1760 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1761 throws com.google.protobuf.InvalidProtocolBufferException { 1762 return PARSER.parseFrom(data, extensionRegistry); 1763 } 1764 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(byte[] data) 1765 throws com.google.protobuf.InvalidProtocolBufferException { 1766 return PARSER.parseFrom(data); 1767 } 1768 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1769 byte[] data, 1770 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1771 throws com.google.protobuf.InvalidProtocolBufferException { 1772 return PARSER.parseFrom(data, extensionRegistry); 1773 } 1774 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(java.io.InputStream input) 1775 throws java.io.IOException { 1776 return PARSER.parseFrom(input); 1777 } 1778 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1779 java.io.InputStream input, 1780 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1781 throws java.io.IOException { 1782 return PARSER.parseFrom(input, extensionRegistry); 1783 } 1784 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(java.io.InputStream input) 1785 throws java.io.IOException { 1786 return PARSER.parseDelimitedFrom(input); 1787 } 1788 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom( 1789 java.io.InputStream input, 1790 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1791 throws java.io.IOException { 1792 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1793 } 1794 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1795 com.google.protobuf.CodedInputStream input) 1796 throws java.io.IOException { 1797 return PARSER.parseFrom(input); 1798 } 1799 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom( 1800 com.google.protobuf.CodedInputStream input, 1801 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1802 throws java.io.IOException { 1803 return PARSER.parseFrom(input, extensionRegistry); 1804 } 1805 1806 public static Builder newBuilder() { return Builder.create(); } 1807 public Builder newBuilderForType() { return newBuilder(); } 1808 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto prototype) { 1809 return newBuilder().mergeFrom(prototype); 1810 } 1811 public Builder toBuilder() { return newBuilder(this); } 1812 1813 @java.lang.Override 1814 protected Builder newBuilderForType( 1815 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1816 Builder builder = new Builder(parent); 1817 return builder; 1818 } 1819 /** 1820 * Protobuf type {@code hadoop.hdfs.SegmentStateProto} 1821 */ 1822 public static final class Builder extends 1823 com.google.protobuf.GeneratedMessage.Builder<Builder> 1824 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder { 1825 public static final com.google.protobuf.Descriptors.Descriptor 1826 getDescriptor() { 1827 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor; 1828 } 1829 1830 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1831 internalGetFieldAccessorTable() { 1832 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable 1833 .ensureFieldAccessorsInitialized( 1834 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class); 1835 } 1836 1837 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder() 1838 private Builder() { 1839 maybeForceBuilderInitialization(); 1840 } 1841 1842 private Builder( 1843 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1844 super(parent); 1845 maybeForceBuilderInitialization(); 1846 } 1847 private void maybeForceBuilderInitialization() { 1848 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1849 } 1850 } 1851 private static Builder create() { 1852 return new Builder(); 1853 } 1854 1855 public Builder clear() { 1856 super.clear(); 1857 startTxId_ = 0L; 1858 bitField0_ = (bitField0_ & ~0x00000001); 1859 endTxId_ = 0L; 1860 bitField0_ = (bitField0_ & ~0x00000002); 1861 isInProgress_ = false; 1862 bitField0_ = (bitField0_ & ~0x00000004); 1863 return this; 1864 } 1865 1866 public Builder clone() { 1867 return create().mergeFrom(buildPartial()); 1868 } 1869 1870 public com.google.protobuf.Descriptors.Descriptor 1871 getDescriptorForType() { 1872 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor; 1873 } 1874 1875 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getDefaultInstanceForType() { 1876 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 1877 } 1878 1879 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto build() { 1880 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = buildPartial(); 1881 if (!result.isInitialized()) { 1882 throw newUninitializedMessageException(result); 1883 } 1884 return result; 1885 } 1886 1887 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto buildPartial() { 1888 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto(this); 1889 int from_bitField0_ = bitField0_; 1890 int to_bitField0_ = 0; 1891 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1892 to_bitField0_ |= 0x00000001; 1893 } 1894 result.startTxId_ = startTxId_; 1895 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 1896 to_bitField0_ |= 0x00000002; 1897 } 1898 result.endTxId_ = endTxId_; 1899 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 1900 to_bitField0_ |= 0x00000004; 1901 } 1902 result.isInProgress_ = isInProgress_; 1903 result.bitField0_ = to_bitField0_; 1904 onBuilt(); 1905 return result; 1906 } 1907 1908 public Builder mergeFrom(com.google.protobuf.Message other) { 1909 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) { 1910 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)other); 1911 } else { 1912 super.mergeFrom(other); 1913 return this; 1914 } 1915 } 1916 1917 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other) { 1918 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) return this; 1919 if (other.hasStartTxId()) { 1920 setStartTxId(other.getStartTxId()); 1921 } 1922 if (other.hasEndTxId()) { 1923 setEndTxId(other.getEndTxId()); 1924 } 1925 if (other.hasIsInProgress()) { 1926 setIsInProgress(other.getIsInProgress()); 1927 } 1928 this.mergeUnknownFields(other.getUnknownFields()); 1929 return this; 1930 } 1931 1932 public final boolean isInitialized() { 1933 if (!hasStartTxId()) { 1934 1935 return false; 1936 } 1937 if (!hasEndTxId()) { 1938 1939 return false; 1940 } 1941 if (!hasIsInProgress()) { 1942 1943 return false; 1944 } 1945 return true; 1946 } 1947 1948 public Builder mergeFrom( 1949 com.google.protobuf.CodedInputStream input, 1950 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1951 throws java.io.IOException { 1952 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parsedMessage = null; 1953 try { 1954 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1955 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1956 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) e.getUnfinishedMessage(); 1957 throw e; 1958 } finally { 1959 if (parsedMessage != null) { 1960 mergeFrom(parsedMessage); 1961 } 1962 } 1963 return this; 1964 } 1965 private int bitField0_; 1966 1967 // required uint64 startTxId = 1; 1968 private long startTxId_ ; 1969 /** 1970 * <code>required uint64 startTxId = 1;</code> 1971 */ 1972 public boolean hasStartTxId() { 1973 return ((bitField0_ & 0x00000001) == 0x00000001); 1974 } 1975 /** 1976 * <code>required uint64 startTxId = 1;</code> 1977 */ 1978 public long getStartTxId() { 1979 return startTxId_; 1980 } 1981 /** 1982 * <code>required uint64 startTxId = 1;</code> 1983 */ 1984 public Builder setStartTxId(long value) { 1985 bitField0_ |= 0x00000001; 1986 startTxId_ = value; 1987 onChanged(); 1988 return this; 1989 } 1990 /** 1991 * <code>required uint64 startTxId = 1;</code> 1992 */ 1993 public Builder clearStartTxId() { 1994 bitField0_ = (bitField0_ & ~0x00000001); 1995 startTxId_ = 0L; 1996 onChanged(); 1997 return this; 1998 } 1999 2000 // required uint64 endTxId = 2; 2001 private long endTxId_ ; 2002 /** 2003 * <code>required uint64 endTxId = 2;</code> 2004 */ 2005 public boolean hasEndTxId() { 2006 return ((bitField0_ & 0x00000002) == 0x00000002); 2007 } 2008 /** 2009 * <code>required uint64 endTxId = 2;</code> 2010 */ 2011 public long getEndTxId() { 2012 return endTxId_; 2013 } 2014 /** 2015 * <code>required uint64 endTxId = 2;</code> 2016 */ 2017 public Builder setEndTxId(long value) { 2018 bitField0_ |= 0x00000002; 2019 endTxId_ = value; 2020 onChanged(); 2021 return this; 2022 } 2023 /** 2024 * <code>required uint64 endTxId = 2;</code> 2025 */ 2026 public Builder clearEndTxId() { 2027 bitField0_ = (bitField0_ & ~0x00000002); 2028 endTxId_ = 0L; 2029 onChanged(); 2030 return this; 2031 } 2032 2033 // required bool isInProgress = 3; 2034 private boolean isInProgress_ ; 2035 /** 2036 * <code>required bool isInProgress = 3;</code> 2037 */ 2038 public boolean hasIsInProgress() { 2039 return ((bitField0_ & 0x00000004) == 0x00000004); 2040 } 2041 /** 2042 * <code>required bool isInProgress = 3;</code> 2043 */ 2044 public boolean getIsInProgress() { 2045 return isInProgress_; 2046 } 2047 /** 2048 * <code>required bool isInProgress = 3;</code> 2049 */ 2050 public Builder setIsInProgress(boolean value) { 2051 bitField0_ |= 0x00000004; 2052 isInProgress_ = value; 2053 onChanged(); 2054 return this; 2055 } 2056 /** 2057 * <code>required bool isInProgress = 3;</code> 2058 */ 2059 public Builder clearIsInProgress() { 2060 bitField0_ = (bitField0_ & ~0x00000004); 2061 isInProgress_ = false; 2062 onChanged(); 2063 return this; 2064 } 2065 2066 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.SegmentStateProto) 2067 } 2068 2069 static { 2070 defaultInstance = new SegmentStateProto(true); 2071 defaultInstance.initFields(); 2072 } 2073 2074 // @@protoc_insertion_point(class_scope:hadoop.hdfs.SegmentStateProto) 2075 } 2076 2077 public interface PersistedRecoveryPaxosDataOrBuilder 2078 extends com.google.protobuf.MessageOrBuilder { 2079 2080 // required .hadoop.hdfs.SegmentStateProto segmentState = 1; 2081 /** 2082 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2083 */ 2084 boolean hasSegmentState(); 2085 /** 2086 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2087 */ 2088 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState(); 2089 /** 2090 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2091 */ 2092 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder(); 2093 2094 // required uint64 acceptedInEpoch = 2; 2095 /** 2096 * <code>required uint64 acceptedInEpoch = 2;</code> 2097 */ 2098 boolean hasAcceptedInEpoch(); 2099 /** 2100 * <code>required uint64 acceptedInEpoch = 2;</code> 2101 */ 2102 long getAcceptedInEpoch(); 2103 } 2104 /** 2105 * Protobuf type {@code hadoop.hdfs.PersistedRecoveryPaxosData} 2106 * 2107 * <pre> 2108 ** 2109 * The storage format used on local disk for previously 2110 * accepted decisions. 2111 * </pre> 2112 */ 2113 public static final class PersistedRecoveryPaxosData extends 2114 com.google.protobuf.GeneratedMessage 2115 implements PersistedRecoveryPaxosDataOrBuilder { 2116 // Use PersistedRecoveryPaxosData.newBuilder() to construct. 2117 private PersistedRecoveryPaxosData(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2118 super(builder); 2119 this.unknownFields = builder.getUnknownFields(); 2120 } 2121 private PersistedRecoveryPaxosData(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2122 2123 private static final PersistedRecoveryPaxosData defaultInstance; 2124 public static PersistedRecoveryPaxosData getDefaultInstance() { 2125 return defaultInstance; 2126 } 2127 2128 public PersistedRecoveryPaxosData getDefaultInstanceForType() { 2129 return defaultInstance; 2130 } 2131 2132 private final com.google.protobuf.UnknownFieldSet unknownFields; 2133 @java.lang.Override 2134 public final com.google.protobuf.UnknownFieldSet 2135 getUnknownFields() { 2136 return this.unknownFields; 2137 } 2138 private PersistedRecoveryPaxosData( 2139 com.google.protobuf.CodedInputStream input, 2140 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2141 throws com.google.protobuf.InvalidProtocolBufferException { 2142 initFields(); 2143 int mutable_bitField0_ = 0; 2144 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2145 com.google.protobuf.UnknownFieldSet.newBuilder(); 2146 try { 2147 boolean done = false; 2148 while (!done) { 2149 int tag = input.readTag(); 2150 switch (tag) { 2151 case 0: 2152 done = true; 2153 break; 2154 default: { 2155 if (!parseUnknownField(input, unknownFields, 2156 extensionRegistry, tag)) { 2157 done = true; 2158 } 2159 break; 2160 } 2161 case 10: { 2162 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null; 2163 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2164 subBuilder = segmentState_.toBuilder(); 2165 } 2166 segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry); 2167 if (subBuilder != null) { 2168 subBuilder.mergeFrom(segmentState_); 2169 segmentState_ = subBuilder.buildPartial(); 2170 } 2171 bitField0_ |= 0x00000001; 2172 break; 2173 } 2174 case 16: { 2175 bitField0_ |= 0x00000002; 2176 acceptedInEpoch_ = input.readUInt64(); 2177 break; 2178 } 2179 } 2180 } 2181 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2182 throw e.setUnfinishedMessage(this); 2183 } catch (java.io.IOException e) { 2184 throw new com.google.protobuf.InvalidProtocolBufferException( 2185 e.getMessage()).setUnfinishedMessage(this); 2186 } finally { 2187 this.unknownFields = unknownFields.build(); 2188 makeExtensionsImmutable(); 2189 } 2190 } 2191 public static final com.google.protobuf.Descriptors.Descriptor 2192 getDescriptor() { 2193 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor; 2194 } 2195 2196 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2197 internalGetFieldAccessorTable() { 2198 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable 2199 .ensureFieldAccessorsInitialized( 2200 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class); 2201 } 2202 2203 public static com.google.protobuf.Parser<PersistedRecoveryPaxosData> PARSER = 2204 new com.google.protobuf.AbstractParser<PersistedRecoveryPaxosData>() { 2205 public PersistedRecoveryPaxosData parsePartialFrom( 2206 com.google.protobuf.CodedInputStream input, 2207 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2208 throws com.google.protobuf.InvalidProtocolBufferException { 2209 return new PersistedRecoveryPaxosData(input, extensionRegistry); 2210 } 2211 }; 2212 2213 @java.lang.Override 2214 public com.google.protobuf.Parser<PersistedRecoveryPaxosData> getParserForType() { 2215 return PARSER; 2216 } 2217 2218 private int bitField0_; 2219 // required .hadoop.hdfs.SegmentStateProto segmentState = 1; 2220 public static final int SEGMENTSTATE_FIELD_NUMBER = 1; 2221 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_; 2222 /** 2223 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2224 */ 2225 public boolean hasSegmentState() { 2226 return ((bitField0_ & 0x00000001) == 0x00000001); 2227 } 2228 /** 2229 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2230 */ 2231 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() { 2232 return segmentState_; 2233 } 2234 /** 2235 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2236 */ 2237 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() { 2238 return segmentState_; 2239 } 2240 2241 // required uint64 acceptedInEpoch = 2; 2242 public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2; 2243 private long acceptedInEpoch_; 2244 /** 2245 * <code>required uint64 acceptedInEpoch = 2;</code> 2246 */ 2247 public boolean hasAcceptedInEpoch() { 2248 return ((bitField0_ & 0x00000002) == 0x00000002); 2249 } 2250 /** 2251 * <code>required uint64 acceptedInEpoch = 2;</code> 2252 */ 2253 public long getAcceptedInEpoch() { 2254 return acceptedInEpoch_; 2255 } 2256 2257 private void initFields() { 2258 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 2259 acceptedInEpoch_ = 0L; 2260 } 2261 private byte memoizedIsInitialized = -1; 2262 public final boolean isInitialized() { 2263 byte isInitialized = memoizedIsInitialized; 2264 if (isInitialized != -1) return isInitialized == 1; 2265 2266 if (!hasSegmentState()) { 2267 memoizedIsInitialized = 0; 2268 return false; 2269 } 2270 if (!hasAcceptedInEpoch()) { 2271 memoizedIsInitialized = 0; 2272 return false; 2273 } 2274 if (!getSegmentState().isInitialized()) { 2275 memoizedIsInitialized = 0; 2276 return false; 2277 } 2278 memoizedIsInitialized = 1; 2279 return true; 2280 } 2281 2282 public void writeTo(com.google.protobuf.CodedOutputStream output) 2283 throws java.io.IOException { 2284 getSerializedSize(); 2285 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2286 output.writeMessage(1, segmentState_); 2287 } 2288 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2289 output.writeUInt64(2, acceptedInEpoch_); 2290 } 2291 getUnknownFields().writeTo(output); 2292 } 2293 2294 private int memoizedSerializedSize = -1; 2295 public int getSerializedSize() { 2296 int size = memoizedSerializedSize; 2297 if (size != -1) return size; 2298 2299 size = 0; 2300 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2301 size += com.google.protobuf.CodedOutputStream 2302 .computeMessageSize(1, segmentState_); 2303 } 2304 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2305 size += com.google.protobuf.CodedOutputStream 2306 .computeUInt64Size(2, acceptedInEpoch_); 2307 } 2308 size += getUnknownFields().getSerializedSize(); 2309 memoizedSerializedSize = size; 2310 return size; 2311 } 2312 2313 private static final long serialVersionUID = 0L; 2314 @java.lang.Override 2315 protected java.lang.Object writeReplace() 2316 throws java.io.ObjectStreamException { 2317 return super.writeReplace(); 2318 } 2319 2320 @java.lang.Override 2321 public boolean equals(final java.lang.Object obj) { 2322 if (obj == this) { 2323 return true; 2324 } 2325 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)) { 2326 return super.equals(obj); 2327 } 2328 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) obj; 2329 2330 boolean result = true; 2331 result = result && (hasSegmentState() == other.hasSegmentState()); 2332 if (hasSegmentState()) { 2333 result = result && getSegmentState() 2334 .equals(other.getSegmentState()); 2335 } 2336 result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch()); 2337 if (hasAcceptedInEpoch()) { 2338 result = result && (getAcceptedInEpoch() 2339 == other.getAcceptedInEpoch()); 2340 } 2341 result = result && 2342 getUnknownFields().equals(other.getUnknownFields()); 2343 return result; 2344 } 2345 2346 private int memoizedHashCode = 0; 2347 @java.lang.Override 2348 public int hashCode() { 2349 if (memoizedHashCode != 0) { 2350 return memoizedHashCode; 2351 } 2352 int hash = 41; 2353 hash = (19 * hash) + getDescriptorForType().hashCode(); 2354 if (hasSegmentState()) { 2355 hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER; 2356 hash = (53 * hash) + getSegmentState().hashCode(); 2357 } 2358 if (hasAcceptedInEpoch()) { 2359 hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER; 2360 hash = (53 * hash) + hashLong(getAcceptedInEpoch()); 2361 } 2362 hash = (29 * hash) + getUnknownFields().hashCode(); 2363 memoizedHashCode = hash; 2364 return hash; 2365 } 2366 2367 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2368 com.google.protobuf.ByteString data) 2369 throws com.google.protobuf.InvalidProtocolBufferException { 2370 return PARSER.parseFrom(data); 2371 } 2372 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2373 com.google.protobuf.ByteString data, 2374 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2375 throws com.google.protobuf.InvalidProtocolBufferException { 2376 return PARSER.parseFrom(data, extensionRegistry); 2377 } 2378 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(byte[] data) 2379 throws com.google.protobuf.InvalidProtocolBufferException { 2380 return PARSER.parseFrom(data); 2381 } 2382 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2383 byte[] data, 2384 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2385 throws com.google.protobuf.InvalidProtocolBufferException { 2386 return PARSER.parseFrom(data, extensionRegistry); 2387 } 2388 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(java.io.InputStream input) 2389 throws java.io.IOException { 2390 return PARSER.parseFrom(input); 2391 } 2392 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2393 java.io.InputStream input, 2394 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2395 throws java.io.IOException { 2396 return PARSER.parseFrom(input, extensionRegistry); 2397 } 2398 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(java.io.InputStream input) 2399 throws java.io.IOException { 2400 return PARSER.parseDelimitedFrom(input); 2401 } 2402 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom( 2403 java.io.InputStream input, 2404 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2405 throws java.io.IOException { 2406 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2407 } 2408 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2409 com.google.protobuf.CodedInputStream input) 2410 throws java.io.IOException { 2411 return PARSER.parseFrom(input); 2412 } 2413 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom( 2414 com.google.protobuf.CodedInputStream input, 2415 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2416 throws java.io.IOException { 2417 return PARSER.parseFrom(input, extensionRegistry); 2418 } 2419 2420 public static Builder newBuilder() { return Builder.create(); } 2421 public Builder newBuilderForType() { return newBuilder(); } 2422 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData prototype) { 2423 return newBuilder().mergeFrom(prototype); 2424 } 2425 public Builder toBuilder() { return newBuilder(this); } 2426 2427 @java.lang.Override 2428 protected Builder newBuilderForType( 2429 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2430 Builder builder = new Builder(parent); 2431 return builder; 2432 } 2433 /** 2434 * Protobuf type {@code hadoop.hdfs.PersistedRecoveryPaxosData} 2435 * 2436 * <pre> 2437 ** 2438 * The storage format used on local disk for previously 2439 * accepted decisions. 2440 * </pre> 2441 */ 2442 public static final class Builder extends 2443 com.google.protobuf.GeneratedMessage.Builder<Builder> 2444 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosDataOrBuilder { 2445 public static final com.google.protobuf.Descriptors.Descriptor 2446 getDescriptor() { 2447 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor; 2448 } 2449 2450 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2451 internalGetFieldAccessorTable() { 2452 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable 2453 .ensureFieldAccessorsInitialized( 2454 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class); 2455 } 2456 2457 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.newBuilder() 2458 private Builder() { 2459 maybeForceBuilderInitialization(); 2460 } 2461 2462 private Builder( 2463 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2464 super(parent); 2465 maybeForceBuilderInitialization(); 2466 } 2467 private void maybeForceBuilderInitialization() { 2468 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2469 getSegmentStateFieldBuilder(); 2470 } 2471 } 2472 private static Builder create() { 2473 return new Builder(); 2474 } 2475 2476 public Builder clear() { 2477 super.clear(); 2478 if (segmentStateBuilder_ == null) { 2479 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 2480 } else { 2481 segmentStateBuilder_.clear(); 2482 } 2483 bitField0_ = (bitField0_ & ~0x00000001); 2484 acceptedInEpoch_ = 0L; 2485 bitField0_ = (bitField0_ & ~0x00000002); 2486 return this; 2487 } 2488 2489 public Builder clone() { 2490 return create().mergeFrom(buildPartial()); 2491 } 2492 2493 public com.google.protobuf.Descriptors.Descriptor 2494 getDescriptorForType() { 2495 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor; 2496 } 2497 2498 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData getDefaultInstanceForType() { 2499 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance(); 2500 } 2501 2502 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData build() { 2503 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = buildPartial(); 2504 if (!result.isInitialized()) { 2505 throw newUninitializedMessageException(result); 2506 } 2507 return result; 2508 } 2509 2510 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData buildPartial() { 2511 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData(this); 2512 int from_bitField0_ = bitField0_; 2513 int to_bitField0_ = 0; 2514 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 2515 to_bitField0_ |= 0x00000001; 2516 } 2517 if (segmentStateBuilder_ == null) { 2518 result.segmentState_ = segmentState_; 2519 } else { 2520 result.segmentState_ = segmentStateBuilder_.build(); 2521 } 2522 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 2523 to_bitField0_ |= 0x00000002; 2524 } 2525 result.acceptedInEpoch_ = acceptedInEpoch_; 2526 result.bitField0_ = to_bitField0_; 2527 onBuilt(); 2528 return result; 2529 } 2530 2531 public Builder mergeFrom(com.google.protobuf.Message other) { 2532 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) { 2533 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)other); 2534 } else { 2535 super.mergeFrom(other); 2536 return this; 2537 } 2538 } 2539 2540 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other) { 2541 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance()) return this; 2542 if (other.hasSegmentState()) { 2543 mergeSegmentState(other.getSegmentState()); 2544 } 2545 if (other.hasAcceptedInEpoch()) { 2546 setAcceptedInEpoch(other.getAcceptedInEpoch()); 2547 } 2548 this.mergeUnknownFields(other.getUnknownFields()); 2549 return this; 2550 } 2551 2552 public final boolean isInitialized() { 2553 if (!hasSegmentState()) { 2554 2555 return false; 2556 } 2557 if (!hasAcceptedInEpoch()) { 2558 2559 return false; 2560 } 2561 if (!getSegmentState().isInitialized()) { 2562 2563 return false; 2564 } 2565 return true; 2566 } 2567 2568 public Builder mergeFrom( 2569 com.google.protobuf.CodedInputStream input, 2570 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2571 throws java.io.IOException { 2572 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parsedMessage = null; 2573 try { 2574 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2575 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2576 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) e.getUnfinishedMessage(); 2577 throw e; 2578 } finally { 2579 if (parsedMessage != null) { 2580 mergeFrom(parsedMessage); 2581 } 2582 } 2583 return this; 2584 } 2585 private int bitField0_; 2586 2587 // required .hadoop.hdfs.SegmentStateProto segmentState = 1; 2588 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 2589 private com.google.protobuf.SingleFieldBuilder< 2590 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_; 2591 /** 2592 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2593 */ 2594 public boolean hasSegmentState() { 2595 return ((bitField0_ & 0x00000001) == 0x00000001); 2596 } 2597 /** 2598 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2599 */ 2600 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() { 2601 if (segmentStateBuilder_ == null) { 2602 return segmentState_; 2603 } else { 2604 return segmentStateBuilder_.getMessage(); 2605 } 2606 } 2607 /** 2608 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2609 */ 2610 public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 2611 if (segmentStateBuilder_ == null) { 2612 if (value == null) { 2613 throw new NullPointerException(); 2614 } 2615 segmentState_ = value; 2616 onChanged(); 2617 } else { 2618 segmentStateBuilder_.setMessage(value); 2619 } 2620 bitField0_ |= 0x00000001; 2621 return this; 2622 } 2623 /** 2624 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2625 */ 2626 public Builder setSegmentState( 2627 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) { 2628 if (segmentStateBuilder_ == null) { 2629 segmentState_ = builderForValue.build(); 2630 onChanged(); 2631 } else { 2632 segmentStateBuilder_.setMessage(builderForValue.build()); 2633 } 2634 bitField0_ |= 0x00000001; 2635 return this; 2636 } 2637 /** 2638 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2639 */ 2640 public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 2641 if (segmentStateBuilder_ == null) { 2642 if (((bitField0_ & 0x00000001) == 0x00000001) && 2643 segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) { 2644 segmentState_ = 2645 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial(); 2646 } else { 2647 segmentState_ = value; 2648 } 2649 onChanged(); 2650 } else { 2651 segmentStateBuilder_.mergeFrom(value); 2652 } 2653 bitField0_ |= 0x00000001; 2654 return this; 2655 } 2656 /** 2657 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2658 */ 2659 public Builder clearSegmentState() { 2660 if (segmentStateBuilder_ == null) { 2661 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 2662 onChanged(); 2663 } else { 2664 segmentStateBuilder_.clear(); 2665 } 2666 bitField0_ = (bitField0_ & ~0x00000001); 2667 return this; 2668 } 2669 /** 2670 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2671 */ 2672 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() { 2673 bitField0_ |= 0x00000001; 2674 onChanged(); 2675 return getSegmentStateFieldBuilder().getBuilder(); 2676 } 2677 /** 2678 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2679 */ 2680 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() { 2681 if (segmentStateBuilder_ != null) { 2682 return segmentStateBuilder_.getMessageOrBuilder(); 2683 } else { 2684 return segmentState_; 2685 } 2686 } 2687 /** 2688 * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 2689 */ 2690 private com.google.protobuf.SingleFieldBuilder< 2691 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 2692 getSegmentStateFieldBuilder() { 2693 if (segmentStateBuilder_ == null) { 2694 segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder< 2695 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>( 2696 segmentState_, 2697 getParentForChildren(), 2698 isClean()); 2699 segmentState_ = null; 2700 } 2701 return segmentStateBuilder_; 2702 } 2703 2704 // required uint64 acceptedInEpoch = 2; 2705 private long acceptedInEpoch_ ; 2706 /** 2707 * <code>required uint64 acceptedInEpoch = 2;</code> 2708 */ 2709 public boolean hasAcceptedInEpoch() { 2710 return ((bitField0_ & 0x00000002) == 0x00000002); 2711 } 2712 /** 2713 * <code>required uint64 acceptedInEpoch = 2;</code> 2714 */ 2715 public long getAcceptedInEpoch() { 2716 return acceptedInEpoch_; 2717 } 2718 /** 2719 * <code>required uint64 acceptedInEpoch = 2;</code> 2720 */ 2721 public Builder setAcceptedInEpoch(long value) { 2722 bitField0_ |= 0x00000002; 2723 acceptedInEpoch_ = value; 2724 onChanged(); 2725 return this; 2726 } 2727 /** 2728 * <code>required uint64 acceptedInEpoch = 2;</code> 2729 */ 2730 public Builder clearAcceptedInEpoch() { 2731 bitField0_ = (bitField0_ & ~0x00000002); 2732 acceptedInEpoch_ = 0L; 2733 onChanged(); 2734 return this; 2735 } 2736 2737 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PersistedRecoveryPaxosData) 2738 } 2739 2740 static { 2741 defaultInstance = new PersistedRecoveryPaxosData(true); 2742 defaultInstance.initFields(); 2743 } 2744 2745 // @@protoc_insertion_point(class_scope:hadoop.hdfs.PersistedRecoveryPaxosData) 2746 } 2747 2748 public interface JournalRequestProtoOrBuilder 2749 extends com.google.protobuf.MessageOrBuilder { 2750 2751 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 2752 /** 2753 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2754 */ 2755 boolean hasReqInfo(); 2756 /** 2757 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2758 */ 2759 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 2760 /** 2761 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2762 */ 2763 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 2764 2765 // required uint64 firstTxnId = 2; 2766 /** 2767 * <code>required uint64 firstTxnId = 2;</code> 2768 */ 2769 boolean hasFirstTxnId(); 2770 /** 2771 * <code>required uint64 firstTxnId = 2;</code> 2772 */ 2773 long getFirstTxnId(); 2774 2775 // required uint32 numTxns = 3; 2776 /** 2777 * <code>required uint32 numTxns = 3;</code> 2778 */ 2779 boolean hasNumTxns(); 2780 /** 2781 * <code>required uint32 numTxns = 3;</code> 2782 */ 2783 int getNumTxns(); 2784 2785 // required bytes records = 4; 2786 /** 2787 * <code>required bytes records = 4;</code> 2788 */ 2789 boolean hasRecords(); 2790 /** 2791 * <code>required bytes records = 4;</code> 2792 */ 2793 com.google.protobuf.ByteString getRecords(); 2794 2795 // required uint64 segmentTxnId = 5; 2796 /** 2797 * <code>required uint64 segmentTxnId = 5;</code> 2798 */ 2799 boolean hasSegmentTxnId(); 2800 /** 2801 * <code>required uint64 segmentTxnId = 5;</code> 2802 */ 2803 long getSegmentTxnId(); 2804 } 2805 /** 2806 * Protobuf type {@code hadoop.hdfs.JournalRequestProto} 2807 */ 2808 public static final class JournalRequestProto extends 2809 com.google.protobuf.GeneratedMessage 2810 implements JournalRequestProtoOrBuilder { 2811 // Use JournalRequestProto.newBuilder() to construct. 2812 private JournalRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2813 super(builder); 2814 this.unknownFields = builder.getUnknownFields(); 2815 } 2816 private JournalRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2817 2818 private static final JournalRequestProto defaultInstance; 2819 public static JournalRequestProto getDefaultInstance() { 2820 return defaultInstance; 2821 } 2822 2823 public JournalRequestProto getDefaultInstanceForType() { 2824 return defaultInstance; 2825 } 2826 2827 private final com.google.protobuf.UnknownFieldSet unknownFields; 2828 @java.lang.Override 2829 public final com.google.protobuf.UnknownFieldSet 2830 getUnknownFields() { 2831 return this.unknownFields; 2832 } 2833 private JournalRequestProto( 2834 com.google.protobuf.CodedInputStream input, 2835 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2836 throws com.google.protobuf.InvalidProtocolBufferException { 2837 initFields(); 2838 int mutable_bitField0_ = 0; 2839 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2840 com.google.protobuf.UnknownFieldSet.newBuilder(); 2841 try { 2842 boolean done = false; 2843 while (!done) { 2844 int tag = input.readTag(); 2845 switch (tag) { 2846 case 0: 2847 done = true; 2848 break; 2849 default: { 2850 if (!parseUnknownField(input, unknownFields, 2851 extensionRegistry, tag)) { 2852 done = true; 2853 } 2854 break; 2855 } 2856 case 10: { 2857 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 2858 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2859 subBuilder = reqInfo_.toBuilder(); 2860 } 2861 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 2862 if (subBuilder != null) { 2863 subBuilder.mergeFrom(reqInfo_); 2864 reqInfo_ = subBuilder.buildPartial(); 2865 } 2866 bitField0_ |= 0x00000001; 2867 break; 2868 } 2869 case 16: { 2870 bitField0_ |= 0x00000002; 2871 firstTxnId_ = input.readUInt64(); 2872 break; 2873 } 2874 case 24: { 2875 bitField0_ |= 0x00000004; 2876 numTxns_ = input.readUInt32(); 2877 break; 2878 } 2879 case 34: { 2880 bitField0_ |= 0x00000008; 2881 records_ = input.readBytes(); 2882 break; 2883 } 2884 case 40: { 2885 bitField0_ |= 0x00000010; 2886 segmentTxnId_ = input.readUInt64(); 2887 break; 2888 } 2889 } 2890 } 2891 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2892 throw e.setUnfinishedMessage(this); 2893 } catch (java.io.IOException e) { 2894 throw new com.google.protobuf.InvalidProtocolBufferException( 2895 e.getMessage()).setUnfinishedMessage(this); 2896 } finally { 2897 this.unknownFields = unknownFields.build(); 2898 makeExtensionsImmutable(); 2899 } 2900 } 2901 public static final com.google.protobuf.Descriptors.Descriptor 2902 getDescriptor() { 2903 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor; 2904 } 2905 2906 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2907 internalGetFieldAccessorTable() { 2908 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable 2909 .ensureFieldAccessorsInitialized( 2910 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class); 2911 } 2912 2913 public static com.google.protobuf.Parser<JournalRequestProto> PARSER = 2914 new com.google.protobuf.AbstractParser<JournalRequestProto>() { 2915 public JournalRequestProto parsePartialFrom( 2916 com.google.protobuf.CodedInputStream input, 2917 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2918 throws com.google.protobuf.InvalidProtocolBufferException { 2919 return new JournalRequestProto(input, extensionRegistry); 2920 } 2921 }; 2922 2923 @java.lang.Override 2924 public com.google.protobuf.Parser<JournalRequestProto> getParserForType() { 2925 return PARSER; 2926 } 2927 2928 private int bitField0_; 2929 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 2930 public static final int REQINFO_FIELD_NUMBER = 1; 2931 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 2932 /** 2933 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2934 */ 2935 public boolean hasReqInfo() { 2936 return ((bitField0_ & 0x00000001) == 0x00000001); 2937 } 2938 /** 2939 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2940 */ 2941 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 2942 return reqInfo_; 2943 } 2944 /** 2945 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 2946 */ 2947 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 2948 return reqInfo_; 2949 } 2950 2951 // required uint64 firstTxnId = 2; 2952 public static final int FIRSTTXNID_FIELD_NUMBER = 2; 2953 private long firstTxnId_; 2954 /** 2955 * <code>required uint64 firstTxnId = 2;</code> 2956 */ 2957 public boolean hasFirstTxnId() { 2958 return ((bitField0_ & 0x00000002) == 0x00000002); 2959 } 2960 /** 2961 * <code>required uint64 firstTxnId = 2;</code> 2962 */ 2963 public long getFirstTxnId() { 2964 return firstTxnId_; 2965 } 2966 2967 // required uint32 numTxns = 3; 2968 public static final int NUMTXNS_FIELD_NUMBER = 3; 2969 private int numTxns_; 2970 /** 2971 * <code>required uint32 numTxns = 3;</code> 2972 */ 2973 public boolean hasNumTxns() { 2974 return ((bitField0_ & 0x00000004) == 0x00000004); 2975 } 2976 /** 2977 * <code>required uint32 numTxns = 3;</code> 2978 */ 2979 public int getNumTxns() { 2980 return numTxns_; 2981 } 2982 2983 // required bytes records = 4; 2984 public static final int RECORDS_FIELD_NUMBER = 4; 2985 private com.google.protobuf.ByteString records_; 2986 /** 2987 * <code>required bytes records = 4;</code> 2988 */ 2989 public boolean hasRecords() { 2990 return ((bitField0_ & 0x00000008) == 0x00000008); 2991 } 2992 /** 2993 * <code>required bytes records = 4;</code> 2994 */ 2995 public com.google.protobuf.ByteString getRecords() { 2996 return records_; 2997 } 2998 2999 // required uint64 segmentTxnId = 5; 3000 public static final int SEGMENTTXNID_FIELD_NUMBER = 5; 3001 private long segmentTxnId_; 3002 /** 3003 * <code>required uint64 segmentTxnId = 5;</code> 3004 */ 3005 public boolean hasSegmentTxnId() { 3006 return ((bitField0_ & 0x00000010) == 0x00000010); 3007 } 3008 /** 3009 * <code>required uint64 segmentTxnId = 5;</code> 3010 */ 3011 public long getSegmentTxnId() { 3012 return segmentTxnId_; 3013 } 3014 3015 private void initFields() { 3016 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 3017 firstTxnId_ = 0L; 3018 numTxns_ = 0; 3019 records_ = com.google.protobuf.ByteString.EMPTY; 3020 segmentTxnId_ = 0L; 3021 } 3022 private byte memoizedIsInitialized = -1; 3023 public final boolean isInitialized() { 3024 byte isInitialized = memoizedIsInitialized; 3025 if (isInitialized != -1) return isInitialized == 1; 3026 3027 if (!hasReqInfo()) { 3028 memoizedIsInitialized = 0; 3029 return false; 3030 } 3031 if (!hasFirstTxnId()) { 3032 memoizedIsInitialized = 0; 3033 return false; 3034 } 3035 if (!hasNumTxns()) { 3036 memoizedIsInitialized = 0; 3037 return false; 3038 } 3039 if (!hasRecords()) { 3040 memoizedIsInitialized = 0; 3041 return false; 3042 } 3043 if (!hasSegmentTxnId()) { 3044 memoizedIsInitialized = 0; 3045 return false; 3046 } 3047 if (!getReqInfo().isInitialized()) { 3048 memoizedIsInitialized = 0; 3049 return false; 3050 } 3051 memoizedIsInitialized = 1; 3052 return true; 3053 } 3054 3055 public void writeTo(com.google.protobuf.CodedOutputStream output) 3056 throws java.io.IOException { 3057 getSerializedSize(); 3058 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3059 output.writeMessage(1, reqInfo_); 3060 } 3061 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3062 output.writeUInt64(2, firstTxnId_); 3063 } 3064 if (((bitField0_ & 0x00000004) == 0x00000004)) { 3065 output.writeUInt32(3, numTxns_); 3066 } 3067 if (((bitField0_ & 0x00000008) == 0x00000008)) { 3068 output.writeBytes(4, records_); 3069 } 3070 if (((bitField0_ & 0x00000010) == 0x00000010)) { 3071 output.writeUInt64(5, segmentTxnId_); 3072 } 3073 getUnknownFields().writeTo(output); 3074 } 3075 3076 private int memoizedSerializedSize = -1; 3077 public int getSerializedSize() { 3078 int size = memoizedSerializedSize; 3079 if (size != -1) return size; 3080 3081 size = 0; 3082 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3083 size += com.google.protobuf.CodedOutputStream 3084 .computeMessageSize(1, reqInfo_); 3085 } 3086 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3087 size += com.google.protobuf.CodedOutputStream 3088 .computeUInt64Size(2, firstTxnId_); 3089 } 3090 if (((bitField0_ & 0x00000004) == 0x00000004)) { 3091 size += com.google.protobuf.CodedOutputStream 3092 .computeUInt32Size(3, numTxns_); 3093 } 3094 if (((bitField0_ & 0x00000008) == 0x00000008)) { 3095 size += com.google.protobuf.CodedOutputStream 3096 .computeBytesSize(4, records_); 3097 } 3098 if (((bitField0_ & 0x00000010) == 0x00000010)) { 3099 size += com.google.protobuf.CodedOutputStream 3100 .computeUInt64Size(5, segmentTxnId_); 3101 } 3102 size += getUnknownFields().getSerializedSize(); 3103 memoizedSerializedSize = size; 3104 return size; 3105 } 3106 3107 private static final long serialVersionUID = 0L; 3108 @java.lang.Override 3109 protected java.lang.Object writeReplace() 3110 throws java.io.ObjectStreamException { 3111 return super.writeReplace(); 3112 } 3113 3114 @java.lang.Override 3115 public boolean equals(final java.lang.Object obj) { 3116 if (obj == this) { 3117 return true; 3118 } 3119 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)) { 3120 return super.equals(obj); 3121 } 3122 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) obj; 3123 3124 boolean result = true; 3125 result = result && (hasReqInfo() == other.hasReqInfo()); 3126 if (hasReqInfo()) { 3127 result = result && getReqInfo() 3128 .equals(other.getReqInfo()); 3129 } 3130 result = result && (hasFirstTxnId() == other.hasFirstTxnId()); 3131 if (hasFirstTxnId()) { 3132 result = result && (getFirstTxnId() 3133 == other.getFirstTxnId()); 3134 } 3135 result = result && (hasNumTxns() == other.hasNumTxns()); 3136 if (hasNumTxns()) { 3137 result = result && (getNumTxns() 3138 == other.getNumTxns()); 3139 } 3140 result = result && (hasRecords() == other.hasRecords()); 3141 if (hasRecords()) { 3142 result = result && getRecords() 3143 .equals(other.getRecords()); 3144 } 3145 result = result && (hasSegmentTxnId() == other.hasSegmentTxnId()); 3146 if (hasSegmentTxnId()) { 3147 result = result && (getSegmentTxnId() 3148 == other.getSegmentTxnId()); 3149 } 3150 result = result && 3151 getUnknownFields().equals(other.getUnknownFields()); 3152 return result; 3153 } 3154 3155 private int memoizedHashCode = 0; 3156 @java.lang.Override 3157 public int hashCode() { 3158 if (memoizedHashCode != 0) { 3159 return memoizedHashCode; 3160 } 3161 int hash = 41; 3162 hash = (19 * hash) + getDescriptorForType().hashCode(); 3163 if (hasReqInfo()) { 3164 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 3165 hash = (53 * hash) + getReqInfo().hashCode(); 3166 } 3167 if (hasFirstTxnId()) { 3168 hash = (37 * hash) + FIRSTTXNID_FIELD_NUMBER; 3169 hash = (53 * hash) + hashLong(getFirstTxnId()); 3170 } 3171 if (hasNumTxns()) { 3172 hash = (37 * hash) + NUMTXNS_FIELD_NUMBER; 3173 hash = (53 * hash) + getNumTxns(); 3174 } 3175 if (hasRecords()) { 3176 hash = (37 * hash) + RECORDS_FIELD_NUMBER; 3177 hash = (53 * hash) + getRecords().hashCode(); 3178 } 3179 if (hasSegmentTxnId()) { 3180 hash = (37 * hash) + SEGMENTTXNID_FIELD_NUMBER; 3181 hash = (53 * hash) + hashLong(getSegmentTxnId()); 3182 } 3183 hash = (29 * hash) + getUnknownFields().hashCode(); 3184 memoizedHashCode = hash; 3185 return hash; 3186 } 3187 3188 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3189 com.google.protobuf.ByteString data) 3190 throws com.google.protobuf.InvalidProtocolBufferException { 3191 return PARSER.parseFrom(data); 3192 } 3193 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3194 com.google.protobuf.ByteString data, 3195 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3196 throws com.google.protobuf.InvalidProtocolBufferException { 3197 return PARSER.parseFrom(data, extensionRegistry); 3198 } 3199 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(byte[] data) 3200 throws com.google.protobuf.InvalidProtocolBufferException { 3201 return PARSER.parseFrom(data); 3202 } 3203 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3204 byte[] data, 3205 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3206 throws com.google.protobuf.InvalidProtocolBufferException { 3207 return PARSER.parseFrom(data, extensionRegistry); 3208 } 3209 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(java.io.InputStream input) 3210 throws java.io.IOException { 3211 return PARSER.parseFrom(input); 3212 } 3213 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3214 java.io.InputStream input, 3215 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3216 throws java.io.IOException { 3217 return PARSER.parseFrom(input, extensionRegistry); 3218 } 3219 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(java.io.InputStream input) 3220 throws java.io.IOException { 3221 return PARSER.parseDelimitedFrom(input); 3222 } 3223 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom( 3224 java.io.InputStream input, 3225 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3226 throws java.io.IOException { 3227 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3228 } 3229 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3230 com.google.protobuf.CodedInputStream input) 3231 throws java.io.IOException { 3232 return PARSER.parseFrom(input); 3233 } 3234 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom( 3235 com.google.protobuf.CodedInputStream input, 3236 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3237 throws java.io.IOException { 3238 return PARSER.parseFrom(input, extensionRegistry); 3239 } 3240 3241 public static Builder newBuilder() { return Builder.create(); } 3242 public Builder newBuilderForType() { return newBuilder(); } 3243 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto prototype) { 3244 return newBuilder().mergeFrom(prototype); 3245 } 3246 public Builder toBuilder() { return newBuilder(this); } 3247 3248 @java.lang.Override 3249 protected Builder newBuilderForType( 3250 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3251 Builder builder = new Builder(parent); 3252 return builder; 3253 } 3254 /** 3255 * Protobuf type {@code hadoop.hdfs.JournalRequestProto} 3256 */ 3257 public static final class Builder extends 3258 com.google.protobuf.GeneratedMessage.Builder<Builder> 3259 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProtoOrBuilder { 3260 public static final com.google.protobuf.Descriptors.Descriptor 3261 getDescriptor() { 3262 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor; 3263 } 3264 3265 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3266 internalGetFieldAccessorTable() { 3267 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable 3268 .ensureFieldAccessorsInitialized( 3269 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class); 3270 } 3271 3272 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.newBuilder() 3273 private Builder() { 3274 maybeForceBuilderInitialization(); 3275 } 3276 3277 private Builder( 3278 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3279 super(parent); 3280 maybeForceBuilderInitialization(); 3281 } 3282 private void maybeForceBuilderInitialization() { 3283 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3284 getReqInfoFieldBuilder(); 3285 } 3286 } 3287 private static Builder create() { 3288 return new Builder(); 3289 } 3290 3291 public Builder clear() { 3292 super.clear(); 3293 if (reqInfoBuilder_ == null) { 3294 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 3295 } else { 3296 reqInfoBuilder_.clear(); 3297 } 3298 bitField0_ = (bitField0_ & ~0x00000001); 3299 firstTxnId_ = 0L; 3300 bitField0_ = (bitField0_ & ~0x00000002); 3301 numTxns_ = 0; 3302 bitField0_ = (bitField0_ & ~0x00000004); 3303 records_ = com.google.protobuf.ByteString.EMPTY; 3304 bitField0_ = (bitField0_ & ~0x00000008); 3305 segmentTxnId_ = 0L; 3306 bitField0_ = (bitField0_ & ~0x00000010); 3307 return this; 3308 } 3309 3310 public Builder clone() { 3311 return create().mergeFrom(buildPartial()); 3312 } 3313 3314 public com.google.protobuf.Descriptors.Descriptor 3315 getDescriptorForType() { 3316 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor; 3317 } 3318 3319 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto getDefaultInstanceForType() { 3320 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance(); 3321 } 3322 3323 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto build() { 3324 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = buildPartial(); 3325 if (!result.isInitialized()) { 3326 throw newUninitializedMessageException(result); 3327 } 3328 return result; 3329 } 3330 3331 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto buildPartial() { 3332 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto(this); 3333 int from_bitField0_ = bitField0_; 3334 int to_bitField0_ = 0; 3335 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 3336 to_bitField0_ |= 0x00000001; 3337 } 3338 if (reqInfoBuilder_ == null) { 3339 result.reqInfo_ = reqInfo_; 3340 } else { 3341 result.reqInfo_ = reqInfoBuilder_.build(); 3342 } 3343 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 3344 to_bitField0_ |= 0x00000002; 3345 } 3346 result.firstTxnId_ = firstTxnId_; 3347 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 3348 to_bitField0_ |= 0x00000004; 3349 } 3350 result.numTxns_ = numTxns_; 3351 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 3352 to_bitField0_ |= 0x00000008; 3353 } 3354 result.records_ = records_; 3355 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 3356 to_bitField0_ |= 0x00000010; 3357 } 3358 result.segmentTxnId_ = segmentTxnId_; 3359 result.bitField0_ = to_bitField0_; 3360 onBuilt(); 3361 return result; 3362 } 3363 3364 public Builder mergeFrom(com.google.protobuf.Message other) { 3365 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) { 3366 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)other); 3367 } else { 3368 super.mergeFrom(other); 3369 return this; 3370 } 3371 } 3372 3373 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other) { 3374 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance()) return this; 3375 if (other.hasReqInfo()) { 3376 mergeReqInfo(other.getReqInfo()); 3377 } 3378 if (other.hasFirstTxnId()) { 3379 setFirstTxnId(other.getFirstTxnId()); 3380 } 3381 if (other.hasNumTxns()) { 3382 setNumTxns(other.getNumTxns()); 3383 } 3384 if (other.hasRecords()) { 3385 setRecords(other.getRecords()); 3386 } 3387 if (other.hasSegmentTxnId()) { 3388 setSegmentTxnId(other.getSegmentTxnId()); 3389 } 3390 this.mergeUnknownFields(other.getUnknownFields()); 3391 return this; 3392 } 3393 3394 public final boolean isInitialized() { 3395 if (!hasReqInfo()) { 3396 3397 return false; 3398 } 3399 if (!hasFirstTxnId()) { 3400 3401 return false; 3402 } 3403 if (!hasNumTxns()) { 3404 3405 return false; 3406 } 3407 if (!hasRecords()) { 3408 3409 return false; 3410 } 3411 if (!hasSegmentTxnId()) { 3412 3413 return false; 3414 } 3415 if (!getReqInfo().isInitialized()) { 3416 3417 return false; 3418 } 3419 return true; 3420 } 3421 3422 public Builder mergeFrom( 3423 com.google.protobuf.CodedInputStream input, 3424 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3425 throws java.io.IOException { 3426 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parsedMessage = null; 3427 try { 3428 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3429 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3430 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) e.getUnfinishedMessage(); 3431 throw e; 3432 } finally { 3433 if (parsedMessage != null) { 3434 mergeFrom(parsedMessage); 3435 } 3436 } 3437 return this; 3438 } 3439 private int bitField0_; 3440 3441 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 3442 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 3443 private com.google.protobuf.SingleFieldBuilder< 3444 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 3445 /** 3446 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3447 */ 3448 public boolean hasReqInfo() { 3449 return ((bitField0_ & 0x00000001) == 0x00000001); 3450 } 3451 /** 3452 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3453 */ 3454 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 3455 if (reqInfoBuilder_ == null) { 3456 return reqInfo_; 3457 } else { 3458 return reqInfoBuilder_.getMessage(); 3459 } 3460 } 3461 /** 3462 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3463 */ 3464 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 3465 if (reqInfoBuilder_ == null) { 3466 if (value == null) { 3467 throw new NullPointerException(); 3468 } 3469 reqInfo_ = value; 3470 onChanged(); 3471 } else { 3472 reqInfoBuilder_.setMessage(value); 3473 } 3474 bitField0_ |= 0x00000001; 3475 return this; 3476 } 3477 /** 3478 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3479 */ 3480 public Builder setReqInfo( 3481 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 3482 if (reqInfoBuilder_ == null) { 3483 reqInfo_ = builderForValue.build(); 3484 onChanged(); 3485 } else { 3486 reqInfoBuilder_.setMessage(builderForValue.build()); 3487 } 3488 bitField0_ |= 0x00000001; 3489 return this; 3490 } 3491 /** 3492 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3493 */ 3494 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 3495 if (reqInfoBuilder_ == null) { 3496 if (((bitField0_ & 0x00000001) == 0x00000001) && 3497 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 3498 reqInfo_ = 3499 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 3500 } else { 3501 reqInfo_ = value; 3502 } 3503 onChanged(); 3504 } else { 3505 reqInfoBuilder_.mergeFrom(value); 3506 } 3507 bitField0_ |= 0x00000001; 3508 return this; 3509 } 3510 /** 3511 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3512 */ 3513 public Builder clearReqInfo() { 3514 if (reqInfoBuilder_ == null) { 3515 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 3516 onChanged(); 3517 } else { 3518 reqInfoBuilder_.clear(); 3519 } 3520 bitField0_ = (bitField0_ & ~0x00000001); 3521 return this; 3522 } 3523 /** 3524 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3525 */ 3526 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 3527 bitField0_ |= 0x00000001; 3528 onChanged(); 3529 return getReqInfoFieldBuilder().getBuilder(); 3530 } 3531 /** 3532 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3533 */ 3534 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 3535 if (reqInfoBuilder_ != null) { 3536 return reqInfoBuilder_.getMessageOrBuilder(); 3537 } else { 3538 return reqInfo_; 3539 } 3540 } 3541 /** 3542 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 3543 */ 3544 private com.google.protobuf.SingleFieldBuilder< 3545 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 3546 getReqInfoFieldBuilder() { 3547 if (reqInfoBuilder_ == null) { 3548 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 3549 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 3550 reqInfo_, 3551 getParentForChildren(), 3552 isClean()); 3553 reqInfo_ = null; 3554 } 3555 return reqInfoBuilder_; 3556 } 3557 3558 // required uint64 firstTxnId = 2; 3559 private long firstTxnId_ ; 3560 /** 3561 * <code>required uint64 firstTxnId = 2;</code> 3562 */ 3563 public boolean hasFirstTxnId() { 3564 return ((bitField0_ & 0x00000002) == 0x00000002); 3565 } 3566 /** 3567 * <code>required uint64 firstTxnId = 2;</code> 3568 */ 3569 public long getFirstTxnId() { 3570 return firstTxnId_; 3571 } 3572 /** 3573 * <code>required uint64 firstTxnId = 2;</code> 3574 */ 3575 public Builder setFirstTxnId(long value) { 3576 bitField0_ |= 0x00000002; 3577 firstTxnId_ = value; 3578 onChanged(); 3579 return this; 3580 } 3581 /** 3582 * <code>required uint64 firstTxnId = 2;</code> 3583 */ 3584 public Builder clearFirstTxnId() { 3585 bitField0_ = (bitField0_ & ~0x00000002); 3586 firstTxnId_ = 0L; 3587 onChanged(); 3588 return this; 3589 } 3590 3591 // required uint32 numTxns = 3; 3592 private int numTxns_ ; 3593 /** 3594 * <code>required uint32 numTxns = 3;</code> 3595 */ 3596 public boolean hasNumTxns() { 3597 return ((bitField0_ & 0x00000004) == 0x00000004); 3598 } 3599 /** 3600 * <code>required uint32 numTxns = 3;</code> 3601 */ 3602 public int getNumTxns() { 3603 return numTxns_; 3604 } 3605 /** 3606 * <code>required uint32 numTxns = 3;</code> 3607 */ 3608 public Builder setNumTxns(int value) { 3609 bitField0_ |= 0x00000004; 3610 numTxns_ = value; 3611 onChanged(); 3612 return this; 3613 } 3614 /** 3615 * <code>required uint32 numTxns = 3;</code> 3616 */ 3617 public Builder clearNumTxns() { 3618 bitField0_ = (bitField0_ & ~0x00000004); 3619 numTxns_ = 0; 3620 onChanged(); 3621 return this; 3622 } 3623 3624 // required bytes records = 4; 3625 private com.google.protobuf.ByteString records_ = com.google.protobuf.ByteString.EMPTY; 3626 /** 3627 * <code>required bytes records = 4;</code> 3628 */ 3629 public boolean hasRecords() { 3630 return ((bitField0_ & 0x00000008) == 0x00000008); 3631 } 3632 /** 3633 * <code>required bytes records = 4;</code> 3634 */ 3635 public com.google.protobuf.ByteString getRecords() { 3636 return records_; 3637 } 3638 /** 3639 * <code>required bytes records = 4;</code> 3640 */ 3641 public Builder setRecords(com.google.protobuf.ByteString value) { 3642 if (value == null) { 3643 throw new NullPointerException(); 3644 } 3645 bitField0_ |= 0x00000008; 3646 records_ = value; 3647 onChanged(); 3648 return this; 3649 } 3650 /** 3651 * <code>required bytes records = 4;</code> 3652 */ 3653 public Builder clearRecords() { 3654 bitField0_ = (bitField0_ & ~0x00000008); 3655 records_ = getDefaultInstance().getRecords(); 3656 onChanged(); 3657 return this; 3658 } 3659 3660 // required uint64 segmentTxnId = 5; 3661 private long segmentTxnId_ ; 3662 /** 3663 * <code>required uint64 segmentTxnId = 5;</code> 3664 */ 3665 public boolean hasSegmentTxnId() { 3666 return ((bitField0_ & 0x00000010) == 0x00000010); 3667 } 3668 /** 3669 * <code>required uint64 segmentTxnId = 5;</code> 3670 */ 3671 public long getSegmentTxnId() { 3672 return segmentTxnId_; 3673 } 3674 /** 3675 * <code>required uint64 segmentTxnId = 5;</code> 3676 */ 3677 public Builder setSegmentTxnId(long value) { 3678 bitField0_ |= 0x00000010; 3679 segmentTxnId_ = value; 3680 onChanged(); 3681 return this; 3682 } 3683 /** 3684 * <code>required uint64 segmentTxnId = 5;</code> 3685 */ 3686 public Builder clearSegmentTxnId() { 3687 bitField0_ = (bitField0_ & ~0x00000010); 3688 segmentTxnId_ = 0L; 3689 onChanged(); 3690 return this; 3691 } 3692 3693 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalRequestProto) 3694 } 3695 3696 static { 3697 defaultInstance = new JournalRequestProto(true); 3698 defaultInstance.initFields(); 3699 } 3700 3701 // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalRequestProto) 3702 } 3703 3704 public interface JournalResponseProtoOrBuilder 3705 extends com.google.protobuf.MessageOrBuilder { 3706 } 3707 /** 3708 * Protobuf type {@code hadoop.hdfs.JournalResponseProto} 3709 */ 3710 public static final class JournalResponseProto extends 3711 com.google.protobuf.GeneratedMessage 3712 implements JournalResponseProtoOrBuilder { 3713 // Use JournalResponseProto.newBuilder() to construct. 3714 private JournalResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3715 super(builder); 3716 this.unknownFields = builder.getUnknownFields(); 3717 } 3718 private JournalResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3719 3720 private static final JournalResponseProto defaultInstance; 3721 public static JournalResponseProto getDefaultInstance() { 3722 return defaultInstance; 3723 } 3724 3725 public JournalResponseProto getDefaultInstanceForType() { 3726 return defaultInstance; 3727 } 3728 3729 private final com.google.protobuf.UnknownFieldSet unknownFields; 3730 @java.lang.Override 3731 public final com.google.protobuf.UnknownFieldSet 3732 getUnknownFields() { 3733 return this.unknownFields; 3734 } 3735 private JournalResponseProto( 3736 com.google.protobuf.CodedInputStream input, 3737 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3738 throws com.google.protobuf.InvalidProtocolBufferException { 3739 initFields(); 3740 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3741 com.google.protobuf.UnknownFieldSet.newBuilder(); 3742 try { 3743 boolean done = false; 3744 while (!done) { 3745 int tag = input.readTag(); 3746 switch (tag) { 3747 case 0: 3748 done = true; 3749 break; 3750 default: { 3751 if (!parseUnknownField(input, unknownFields, 3752 extensionRegistry, tag)) { 3753 done = true; 3754 } 3755 break; 3756 } 3757 } 3758 } 3759 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3760 throw e.setUnfinishedMessage(this); 3761 } catch (java.io.IOException e) { 3762 throw new com.google.protobuf.InvalidProtocolBufferException( 3763 e.getMessage()).setUnfinishedMessage(this); 3764 } finally { 3765 this.unknownFields = unknownFields.build(); 3766 makeExtensionsImmutable(); 3767 } 3768 } 3769 public static final com.google.protobuf.Descriptors.Descriptor 3770 getDescriptor() { 3771 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor; 3772 } 3773 3774 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3775 internalGetFieldAccessorTable() { 3776 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable 3777 .ensureFieldAccessorsInitialized( 3778 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class); 3779 } 3780 3781 public static com.google.protobuf.Parser<JournalResponseProto> PARSER = 3782 new com.google.protobuf.AbstractParser<JournalResponseProto>() { 3783 public JournalResponseProto parsePartialFrom( 3784 com.google.protobuf.CodedInputStream input, 3785 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3786 throws com.google.protobuf.InvalidProtocolBufferException { 3787 return new JournalResponseProto(input, extensionRegistry); 3788 } 3789 }; 3790 3791 @java.lang.Override 3792 public com.google.protobuf.Parser<JournalResponseProto> getParserForType() { 3793 return PARSER; 3794 } 3795 3796 private void initFields() { 3797 } 3798 private byte memoizedIsInitialized = -1; 3799 public final boolean isInitialized() { 3800 byte isInitialized = memoizedIsInitialized; 3801 if (isInitialized != -1) return isInitialized == 1; 3802 3803 memoizedIsInitialized = 1; 3804 return true; 3805 } 3806 3807 public void writeTo(com.google.protobuf.CodedOutputStream output) 3808 throws java.io.IOException { 3809 getSerializedSize(); 3810 getUnknownFields().writeTo(output); 3811 } 3812 3813 private int memoizedSerializedSize = -1; 3814 public int getSerializedSize() { 3815 int size = memoizedSerializedSize; 3816 if (size != -1) return size; 3817 3818 size = 0; 3819 size += getUnknownFields().getSerializedSize(); 3820 memoizedSerializedSize = size; 3821 return size; 3822 } 3823 3824 private static final long serialVersionUID = 0L; 3825 @java.lang.Override 3826 protected java.lang.Object writeReplace() 3827 throws java.io.ObjectStreamException { 3828 return super.writeReplace(); 3829 } 3830 3831 @java.lang.Override 3832 public boolean equals(final java.lang.Object obj) { 3833 if (obj == this) { 3834 return true; 3835 } 3836 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)) { 3837 return super.equals(obj); 3838 } 3839 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) obj; 3840 3841 boolean result = true; 3842 result = result && 3843 getUnknownFields().equals(other.getUnknownFields()); 3844 return result; 3845 } 3846 3847 private int memoizedHashCode = 0; 3848 @java.lang.Override 3849 public int hashCode() { 3850 if (memoizedHashCode != 0) { 3851 return memoizedHashCode; 3852 } 3853 int hash = 41; 3854 hash = (19 * hash) + getDescriptorForType().hashCode(); 3855 hash = (29 * hash) + getUnknownFields().hashCode(); 3856 memoizedHashCode = hash; 3857 return hash; 3858 } 3859 3860 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3861 com.google.protobuf.ByteString data) 3862 throws com.google.protobuf.InvalidProtocolBufferException { 3863 return PARSER.parseFrom(data); 3864 } 3865 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3866 com.google.protobuf.ByteString data, 3867 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3868 throws com.google.protobuf.InvalidProtocolBufferException { 3869 return PARSER.parseFrom(data, extensionRegistry); 3870 } 3871 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(byte[] data) 3872 throws com.google.protobuf.InvalidProtocolBufferException { 3873 return PARSER.parseFrom(data); 3874 } 3875 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3876 byte[] data, 3877 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3878 throws com.google.protobuf.InvalidProtocolBufferException { 3879 return PARSER.parseFrom(data, extensionRegistry); 3880 } 3881 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(java.io.InputStream input) 3882 throws java.io.IOException { 3883 return PARSER.parseFrom(input); 3884 } 3885 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3886 java.io.InputStream input, 3887 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3888 throws java.io.IOException { 3889 return PARSER.parseFrom(input, extensionRegistry); 3890 } 3891 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(java.io.InputStream input) 3892 throws java.io.IOException { 3893 return PARSER.parseDelimitedFrom(input); 3894 } 3895 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom( 3896 java.io.InputStream input, 3897 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3898 throws java.io.IOException { 3899 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3900 } 3901 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3902 com.google.protobuf.CodedInputStream input) 3903 throws java.io.IOException { 3904 return PARSER.parseFrom(input); 3905 } 3906 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom( 3907 com.google.protobuf.CodedInputStream input, 3908 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3909 throws java.io.IOException { 3910 return PARSER.parseFrom(input, extensionRegistry); 3911 } 3912 3913 public static Builder newBuilder() { return Builder.create(); } 3914 public Builder newBuilderForType() { return newBuilder(); } 3915 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto prototype) { 3916 return newBuilder().mergeFrom(prototype); 3917 } 3918 public Builder toBuilder() { return newBuilder(this); } 3919 3920 @java.lang.Override 3921 protected Builder newBuilderForType( 3922 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3923 Builder builder = new Builder(parent); 3924 return builder; 3925 } 3926 /** 3927 * Protobuf type {@code hadoop.hdfs.JournalResponseProto} 3928 */ 3929 public static final class Builder extends 3930 com.google.protobuf.GeneratedMessage.Builder<Builder> 3931 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProtoOrBuilder { 3932 public static final com.google.protobuf.Descriptors.Descriptor 3933 getDescriptor() { 3934 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor; 3935 } 3936 3937 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3938 internalGetFieldAccessorTable() { 3939 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable 3940 .ensureFieldAccessorsInitialized( 3941 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class); 3942 } 3943 3944 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.newBuilder() 3945 private Builder() { 3946 maybeForceBuilderInitialization(); 3947 } 3948 3949 private Builder( 3950 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3951 super(parent); 3952 maybeForceBuilderInitialization(); 3953 } 3954 private void maybeForceBuilderInitialization() { 3955 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3956 } 3957 } 3958 private static Builder create() { 3959 return new Builder(); 3960 } 3961 3962 public Builder clear() { 3963 super.clear(); 3964 return this; 3965 } 3966 3967 public Builder clone() { 3968 return create().mergeFrom(buildPartial()); 3969 } 3970 3971 public com.google.protobuf.Descriptors.Descriptor 3972 getDescriptorForType() { 3973 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor; 3974 } 3975 3976 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto getDefaultInstanceForType() { 3977 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(); 3978 } 3979 3980 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto build() { 3981 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = buildPartial(); 3982 if (!result.isInitialized()) { 3983 throw newUninitializedMessageException(result); 3984 } 3985 return result; 3986 } 3987 3988 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto buildPartial() { 3989 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto(this); 3990 onBuilt(); 3991 return result; 3992 } 3993 3994 public Builder mergeFrom(com.google.protobuf.Message other) { 3995 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) { 3996 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)other); 3997 } else { 3998 super.mergeFrom(other); 3999 return this; 4000 } 4001 } 4002 4003 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other) { 4004 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()) return this; 4005 this.mergeUnknownFields(other.getUnknownFields()); 4006 return this; 4007 } 4008 4009 public final boolean isInitialized() { 4010 return true; 4011 } 4012 4013 public Builder mergeFrom( 4014 com.google.protobuf.CodedInputStream input, 4015 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4016 throws java.io.IOException { 4017 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parsedMessage = null; 4018 try { 4019 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4020 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4021 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) e.getUnfinishedMessage(); 4022 throw e; 4023 } finally { 4024 if (parsedMessage != null) { 4025 mergeFrom(parsedMessage); 4026 } 4027 } 4028 return this; 4029 } 4030 4031 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalResponseProto) 4032 } 4033 4034 static { 4035 defaultInstance = new JournalResponseProto(true); 4036 defaultInstance.initFields(); 4037 } 4038 4039 // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalResponseProto) 4040 } 4041 4042 public interface HeartbeatRequestProtoOrBuilder 4043 extends com.google.protobuf.MessageOrBuilder { 4044 4045 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 4046 /** 4047 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4048 */ 4049 boolean hasReqInfo(); 4050 /** 4051 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4052 */ 4053 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 4054 /** 4055 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4056 */ 4057 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 4058 } 4059 /** 4060 * Protobuf type {@code hadoop.hdfs.HeartbeatRequestProto} 4061 */ 4062 public static final class HeartbeatRequestProto extends 4063 com.google.protobuf.GeneratedMessage 4064 implements HeartbeatRequestProtoOrBuilder { 4065 // Use HeartbeatRequestProto.newBuilder() to construct. 4066 private HeartbeatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4067 super(builder); 4068 this.unknownFields = builder.getUnknownFields(); 4069 } 4070 private HeartbeatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4071 4072 private static final HeartbeatRequestProto defaultInstance; 4073 public static HeartbeatRequestProto getDefaultInstance() { 4074 return defaultInstance; 4075 } 4076 4077 public HeartbeatRequestProto getDefaultInstanceForType() { 4078 return defaultInstance; 4079 } 4080 4081 private final com.google.protobuf.UnknownFieldSet unknownFields; 4082 @java.lang.Override 4083 public final com.google.protobuf.UnknownFieldSet 4084 getUnknownFields() { 4085 return this.unknownFields; 4086 } 4087 private HeartbeatRequestProto( 4088 com.google.protobuf.CodedInputStream input, 4089 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4090 throws com.google.protobuf.InvalidProtocolBufferException { 4091 initFields(); 4092 int mutable_bitField0_ = 0; 4093 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4094 com.google.protobuf.UnknownFieldSet.newBuilder(); 4095 try { 4096 boolean done = false; 4097 while (!done) { 4098 int tag = input.readTag(); 4099 switch (tag) { 4100 case 0: 4101 done = true; 4102 break; 4103 default: { 4104 if (!parseUnknownField(input, unknownFields, 4105 extensionRegistry, tag)) { 4106 done = true; 4107 } 4108 break; 4109 } 4110 case 10: { 4111 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 4112 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4113 subBuilder = reqInfo_.toBuilder(); 4114 } 4115 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 4116 if (subBuilder != null) { 4117 subBuilder.mergeFrom(reqInfo_); 4118 reqInfo_ = subBuilder.buildPartial(); 4119 } 4120 bitField0_ |= 0x00000001; 4121 break; 4122 } 4123 } 4124 } 4125 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4126 throw e.setUnfinishedMessage(this); 4127 } catch (java.io.IOException e) { 4128 throw new com.google.protobuf.InvalidProtocolBufferException( 4129 e.getMessage()).setUnfinishedMessage(this); 4130 } finally { 4131 this.unknownFields = unknownFields.build(); 4132 makeExtensionsImmutable(); 4133 } 4134 } 4135 public static final com.google.protobuf.Descriptors.Descriptor 4136 getDescriptor() { 4137 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor; 4138 } 4139 4140 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4141 internalGetFieldAccessorTable() { 4142 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable 4143 .ensureFieldAccessorsInitialized( 4144 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class); 4145 } 4146 4147 public static com.google.protobuf.Parser<HeartbeatRequestProto> PARSER = 4148 new com.google.protobuf.AbstractParser<HeartbeatRequestProto>() { 4149 public HeartbeatRequestProto parsePartialFrom( 4150 com.google.protobuf.CodedInputStream input, 4151 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4152 throws com.google.protobuf.InvalidProtocolBufferException { 4153 return new HeartbeatRequestProto(input, extensionRegistry); 4154 } 4155 }; 4156 4157 @java.lang.Override 4158 public com.google.protobuf.Parser<HeartbeatRequestProto> getParserForType() { 4159 return PARSER; 4160 } 4161 4162 private int bitField0_; 4163 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 4164 public static final int REQINFO_FIELD_NUMBER = 1; 4165 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 4166 /** 4167 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4168 */ 4169 public boolean hasReqInfo() { 4170 return ((bitField0_ & 0x00000001) == 0x00000001); 4171 } 4172 /** 4173 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4174 */ 4175 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 4176 return reqInfo_; 4177 } 4178 /** 4179 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4180 */ 4181 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 4182 return reqInfo_; 4183 } 4184 4185 private void initFields() { 4186 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 4187 } 4188 private byte memoizedIsInitialized = -1; 4189 public final boolean isInitialized() { 4190 byte isInitialized = memoizedIsInitialized; 4191 if (isInitialized != -1) return isInitialized == 1; 4192 4193 if (!hasReqInfo()) { 4194 memoizedIsInitialized = 0; 4195 return false; 4196 } 4197 if (!getReqInfo().isInitialized()) { 4198 memoizedIsInitialized = 0; 4199 return false; 4200 } 4201 memoizedIsInitialized = 1; 4202 return true; 4203 } 4204 4205 public void writeTo(com.google.protobuf.CodedOutputStream output) 4206 throws java.io.IOException { 4207 getSerializedSize(); 4208 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4209 output.writeMessage(1, reqInfo_); 4210 } 4211 getUnknownFields().writeTo(output); 4212 } 4213 4214 private int memoizedSerializedSize = -1; 4215 public int getSerializedSize() { 4216 int size = memoizedSerializedSize; 4217 if (size != -1) return size; 4218 4219 size = 0; 4220 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4221 size += com.google.protobuf.CodedOutputStream 4222 .computeMessageSize(1, reqInfo_); 4223 } 4224 size += getUnknownFields().getSerializedSize(); 4225 memoizedSerializedSize = size; 4226 return size; 4227 } 4228 4229 private static final long serialVersionUID = 0L; 4230 @java.lang.Override 4231 protected java.lang.Object writeReplace() 4232 throws java.io.ObjectStreamException { 4233 return super.writeReplace(); 4234 } 4235 4236 @java.lang.Override 4237 public boolean equals(final java.lang.Object obj) { 4238 if (obj == this) { 4239 return true; 4240 } 4241 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)) { 4242 return super.equals(obj); 4243 } 4244 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) obj; 4245 4246 boolean result = true; 4247 result = result && (hasReqInfo() == other.hasReqInfo()); 4248 if (hasReqInfo()) { 4249 result = result && getReqInfo() 4250 .equals(other.getReqInfo()); 4251 } 4252 result = result && 4253 getUnknownFields().equals(other.getUnknownFields()); 4254 return result; 4255 } 4256 4257 private int memoizedHashCode = 0; 4258 @java.lang.Override 4259 public int hashCode() { 4260 if (memoizedHashCode != 0) { 4261 return memoizedHashCode; 4262 } 4263 int hash = 41; 4264 hash = (19 * hash) + getDescriptorForType().hashCode(); 4265 if (hasReqInfo()) { 4266 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 4267 hash = (53 * hash) + getReqInfo().hashCode(); 4268 } 4269 hash = (29 * hash) + getUnknownFields().hashCode(); 4270 memoizedHashCode = hash; 4271 return hash; 4272 } 4273 4274 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4275 com.google.protobuf.ByteString data) 4276 throws com.google.protobuf.InvalidProtocolBufferException { 4277 return PARSER.parseFrom(data); 4278 } 4279 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4280 com.google.protobuf.ByteString data, 4281 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4282 throws com.google.protobuf.InvalidProtocolBufferException { 4283 return PARSER.parseFrom(data, extensionRegistry); 4284 } 4285 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(byte[] data) 4286 throws com.google.protobuf.InvalidProtocolBufferException { 4287 return PARSER.parseFrom(data); 4288 } 4289 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4290 byte[] data, 4291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4292 throws com.google.protobuf.InvalidProtocolBufferException { 4293 return PARSER.parseFrom(data, extensionRegistry); 4294 } 4295 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(java.io.InputStream input) 4296 throws java.io.IOException { 4297 return PARSER.parseFrom(input); 4298 } 4299 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4300 java.io.InputStream input, 4301 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4302 throws java.io.IOException { 4303 return PARSER.parseFrom(input, extensionRegistry); 4304 } 4305 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input) 4306 throws java.io.IOException { 4307 return PARSER.parseDelimitedFrom(input); 4308 } 4309 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom( 4310 java.io.InputStream input, 4311 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4312 throws java.io.IOException { 4313 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4314 } 4315 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4316 com.google.protobuf.CodedInputStream input) 4317 throws java.io.IOException { 4318 return PARSER.parseFrom(input); 4319 } 4320 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom( 4321 com.google.protobuf.CodedInputStream input, 4322 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4323 throws java.io.IOException { 4324 return PARSER.parseFrom(input, extensionRegistry); 4325 } 4326 4327 public static Builder newBuilder() { return Builder.create(); } 4328 public Builder newBuilderForType() { return newBuilder(); } 4329 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto prototype) { 4330 return newBuilder().mergeFrom(prototype); 4331 } 4332 public Builder toBuilder() { return newBuilder(this); } 4333 4334 @java.lang.Override 4335 protected Builder newBuilderForType( 4336 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4337 Builder builder = new Builder(parent); 4338 return builder; 4339 } 4340 /** 4341 * Protobuf type {@code hadoop.hdfs.HeartbeatRequestProto} 4342 */ 4343 public static final class Builder extends 4344 com.google.protobuf.GeneratedMessage.Builder<Builder> 4345 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProtoOrBuilder { 4346 public static final com.google.protobuf.Descriptors.Descriptor 4347 getDescriptor() { 4348 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor; 4349 } 4350 4351 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4352 internalGetFieldAccessorTable() { 4353 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable 4354 .ensureFieldAccessorsInitialized( 4355 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class); 4356 } 4357 4358 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.newBuilder() 4359 private Builder() { 4360 maybeForceBuilderInitialization(); 4361 } 4362 4363 private Builder( 4364 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4365 super(parent); 4366 maybeForceBuilderInitialization(); 4367 } 4368 private void maybeForceBuilderInitialization() { 4369 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4370 getReqInfoFieldBuilder(); 4371 } 4372 } 4373 private static Builder create() { 4374 return new Builder(); 4375 } 4376 4377 public Builder clear() { 4378 super.clear(); 4379 if (reqInfoBuilder_ == null) { 4380 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 4381 } else { 4382 reqInfoBuilder_.clear(); 4383 } 4384 bitField0_ = (bitField0_ & ~0x00000001); 4385 return this; 4386 } 4387 4388 public Builder clone() { 4389 return create().mergeFrom(buildPartial()); 4390 } 4391 4392 public com.google.protobuf.Descriptors.Descriptor 4393 getDescriptorForType() { 4394 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor; 4395 } 4396 4397 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() { 4398 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); 4399 } 4400 4401 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto build() { 4402 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = buildPartial(); 4403 if (!result.isInitialized()) { 4404 throw newUninitializedMessageException(result); 4405 } 4406 return result; 4407 } 4408 4409 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto buildPartial() { 4410 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto(this); 4411 int from_bitField0_ = bitField0_; 4412 int to_bitField0_ = 0; 4413 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 4414 to_bitField0_ |= 0x00000001; 4415 } 4416 if (reqInfoBuilder_ == null) { 4417 result.reqInfo_ = reqInfo_; 4418 } else { 4419 result.reqInfo_ = reqInfoBuilder_.build(); 4420 } 4421 result.bitField0_ = to_bitField0_; 4422 onBuilt(); 4423 return result; 4424 } 4425 4426 public Builder mergeFrom(com.google.protobuf.Message other) { 4427 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) { 4428 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)other); 4429 } else { 4430 super.mergeFrom(other); 4431 return this; 4432 } 4433 } 4434 4435 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other) { 4436 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance()) return this; 4437 if (other.hasReqInfo()) { 4438 mergeReqInfo(other.getReqInfo()); 4439 } 4440 this.mergeUnknownFields(other.getUnknownFields()); 4441 return this; 4442 } 4443 4444 public final boolean isInitialized() { 4445 if (!hasReqInfo()) { 4446 4447 return false; 4448 } 4449 if (!getReqInfo().isInitialized()) { 4450 4451 return false; 4452 } 4453 return true; 4454 } 4455 4456 public Builder mergeFrom( 4457 com.google.protobuf.CodedInputStream input, 4458 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4459 throws java.io.IOException { 4460 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parsedMessage = null; 4461 try { 4462 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4463 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4464 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) e.getUnfinishedMessage(); 4465 throw e; 4466 } finally { 4467 if (parsedMessage != null) { 4468 mergeFrom(parsedMessage); 4469 } 4470 } 4471 return this; 4472 } 4473 private int bitField0_; 4474 4475 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 4476 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 4477 private com.google.protobuf.SingleFieldBuilder< 4478 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 4479 /** 4480 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4481 */ 4482 public boolean hasReqInfo() { 4483 return ((bitField0_ & 0x00000001) == 0x00000001); 4484 } 4485 /** 4486 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4487 */ 4488 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 4489 if (reqInfoBuilder_ == null) { 4490 return reqInfo_; 4491 } else { 4492 return reqInfoBuilder_.getMessage(); 4493 } 4494 } 4495 /** 4496 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4497 */ 4498 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 4499 if (reqInfoBuilder_ == null) { 4500 if (value == null) { 4501 throw new NullPointerException(); 4502 } 4503 reqInfo_ = value; 4504 onChanged(); 4505 } else { 4506 reqInfoBuilder_.setMessage(value); 4507 } 4508 bitField0_ |= 0x00000001; 4509 return this; 4510 } 4511 /** 4512 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4513 */ 4514 public Builder setReqInfo( 4515 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 4516 if (reqInfoBuilder_ == null) { 4517 reqInfo_ = builderForValue.build(); 4518 onChanged(); 4519 } else { 4520 reqInfoBuilder_.setMessage(builderForValue.build()); 4521 } 4522 bitField0_ |= 0x00000001; 4523 return this; 4524 } 4525 /** 4526 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4527 */ 4528 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 4529 if (reqInfoBuilder_ == null) { 4530 if (((bitField0_ & 0x00000001) == 0x00000001) && 4531 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 4532 reqInfo_ = 4533 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 4534 } else { 4535 reqInfo_ = value; 4536 } 4537 onChanged(); 4538 } else { 4539 reqInfoBuilder_.mergeFrom(value); 4540 } 4541 bitField0_ |= 0x00000001; 4542 return this; 4543 } 4544 /** 4545 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4546 */ 4547 public Builder clearReqInfo() { 4548 if (reqInfoBuilder_ == null) { 4549 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 4550 onChanged(); 4551 } else { 4552 reqInfoBuilder_.clear(); 4553 } 4554 bitField0_ = (bitField0_ & ~0x00000001); 4555 return this; 4556 } 4557 /** 4558 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4559 */ 4560 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 4561 bitField0_ |= 0x00000001; 4562 onChanged(); 4563 return getReqInfoFieldBuilder().getBuilder(); 4564 } 4565 /** 4566 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4567 */ 4568 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 4569 if (reqInfoBuilder_ != null) { 4570 return reqInfoBuilder_.getMessageOrBuilder(); 4571 } else { 4572 return reqInfo_; 4573 } 4574 } 4575 /** 4576 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4577 */ 4578 private com.google.protobuf.SingleFieldBuilder< 4579 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 4580 getReqInfoFieldBuilder() { 4581 if (reqInfoBuilder_ == null) { 4582 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 4583 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 4584 reqInfo_, 4585 getParentForChildren(), 4586 isClean()); 4587 reqInfo_ = null; 4588 } 4589 return reqInfoBuilder_; 4590 } 4591 4592 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatRequestProto) 4593 } 4594 4595 static { 4596 defaultInstance = new HeartbeatRequestProto(true); 4597 defaultInstance.initFields(); 4598 } 4599 4600 // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatRequestProto) 4601 } 4602 4603 public interface HeartbeatResponseProtoOrBuilder 4604 extends com.google.protobuf.MessageOrBuilder { 4605 } 4606 /** 4607 * Protobuf type {@code hadoop.hdfs.HeartbeatResponseProto} 4608 * 4609 * <pre> 4610 * void response 4611 * </pre> 4612 */ 4613 public static final class HeartbeatResponseProto extends 4614 com.google.protobuf.GeneratedMessage 4615 implements HeartbeatResponseProtoOrBuilder { 4616 // Use HeartbeatResponseProto.newBuilder() to construct. 4617 private HeartbeatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4618 super(builder); 4619 this.unknownFields = builder.getUnknownFields(); 4620 } 4621 private HeartbeatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4622 4623 private static final HeartbeatResponseProto defaultInstance; 4624 public static HeartbeatResponseProto getDefaultInstance() { 4625 return defaultInstance; 4626 } 4627 4628 public HeartbeatResponseProto getDefaultInstanceForType() { 4629 return defaultInstance; 4630 } 4631 4632 private final com.google.protobuf.UnknownFieldSet unknownFields; 4633 @java.lang.Override 4634 public final com.google.protobuf.UnknownFieldSet 4635 getUnknownFields() { 4636 return this.unknownFields; 4637 } 4638 private HeartbeatResponseProto( 4639 com.google.protobuf.CodedInputStream input, 4640 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4641 throws com.google.protobuf.InvalidProtocolBufferException { 4642 initFields(); 4643 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4644 com.google.protobuf.UnknownFieldSet.newBuilder(); 4645 try { 4646 boolean done = false; 4647 while (!done) { 4648 int tag = input.readTag(); 4649 switch (tag) { 4650 case 0: 4651 done = true; 4652 break; 4653 default: { 4654 if (!parseUnknownField(input, unknownFields, 4655 extensionRegistry, tag)) { 4656 done = true; 4657 } 4658 break; 4659 } 4660 } 4661 } 4662 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4663 throw e.setUnfinishedMessage(this); 4664 } catch (java.io.IOException e) { 4665 throw new com.google.protobuf.InvalidProtocolBufferException( 4666 e.getMessage()).setUnfinishedMessage(this); 4667 } finally { 4668 this.unknownFields = unknownFields.build(); 4669 makeExtensionsImmutable(); 4670 } 4671 } 4672 public static final com.google.protobuf.Descriptors.Descriptor 4673 getDescriptor() { 4674 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor; 4675 } 4676 4677 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4678 internalGetFieldAccessorTable() { 4679 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable 4680 .ensureFieldAccessorsInitialized( 4681 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class); 4682 } 4683 4684 public static com.google.protobuf.Parser<HeartbeatResponseProto> PARSER = 4685 new com.google.protobuf.AbstractParser<HeartbeatResponseProto>() { 4686 public HeartbeatResponseProto parsePartialFrom( 4687 com.google.protobuf.CodedInputStream input, 4688 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4689 throws com.google.protobuf.InvalidProtocolBufferException { 4690 return new HeartbeatResponseProto(input, extensionRegistry); 4691 } 4692 }; 4693 4694 @java.lang.Override 4695 public com.google.protobuf.Parser<HeartbeatResponseProto> getParserForType() { 4696 return PARSER; 4697 } 4698 4699 private void initFields() { 4700 } 4701 private byte memoizedIsInitialized = -1; 4702 public final boolean isInitialized() { 4703 byte isInitialized = memoizedIsInitialized; 4704 if (isInitialized != -1) return isInitialized == 1; 4705 4706 memoizedIsInitialized = 1; 4707 return true; 4708 } 4709 4710 public void writeTo(com.google.protobuf.CodedOutputStream output) 4711 throws java.io.IOException { 4712 getSerializedSize(); 4713 getUnknownFields().writeTo(output); 4714 } 4715 4716 private int memoizedSerializedSize = -1; 4717 public int getSerializedSize() { 4718 int size = memoizedSerializedSize; 4719 if (size != -1) return size; 4720 4721 size = 0; 4722 size += getUnknownFields().getSerializedSize(); 4723 memoizedSerializedSize = size; 4724 return size; 4725 } 4726 4727 private static final long serialVersionUID = 0L; 4728 @java.lang.Override 4729 protected java.lang.Object writeReplace() 4730 throws java.io.ObjectStreamException { 4731 return super.writeReplace(); 4732 } 4733 4734 @java.lang.Override 4735 public boolean equals(final java.lang.Object obj) { 4736 if (obj == this) { 4737 return true; 4738 } 4739 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)) { 4740 return super.equals(obj); 4741 } 4742 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) obj; 4743 4744 boolean result = true; 4745 result = result && 4746 getUnknownFields().equals(other.getUnknownFields()); 4747 return result; 4748 } 4749 4750 private int memoizedHashCode = 0; 4751 @java.lang.Override 4752 public int hashCode() { 4753 if (memoizedHashCode != 0) { 4754 return memoizedHashCode; 4755 } 4756 int hash = 41; 4757 hash = (19 * hash) + getDescriptorForType().hashCode(); 4758 hash = (29 * hash) + getUnknownFields().hashCode(); 4759 memoizedHashCode = hash; 4760 return hash; 4761 } 4762 4763 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4764 com.google.protobuf.ByteString data) 4765 throws com.google.protobuf.InvalidProtocolBufferException { 4766 return PARSER.parseFrom(data); 4767 } 4768 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4769 com.google.protobuf.ByteString data, 4770 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4771 throws com.google.protobuf.InvalidProtocolBufferException { 4772 return PARSER.parseFrom(data, extensionRegistry); 4773 } 4774 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(byte[] data) 4775 throws com.google.protobuf.InvalidProtocolBufferException { 4776 return PARSER.parseFrom(data); 4777 } 4778 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4779 byte[] data, 4780 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4781 throws com.google.protobuf.InvalidProtocolBufferException { 4782 return PARSER.parseFrom(data, extensionRegistry); 4783 } 4784 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(java.io.InputStream input) 4785 throws java.io.IOException { 4786 return PARSER.parseFrom(input); 4787 } 4788 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4789 java.io.InputStream input, 4790 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4791 throws java.io.IOException { 4792 return PARSER.parseFrom(input, extensionRegistry); 4793 } 4794 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input) 4795 throws java.io.IOException { 4796 return PARSER.parseDelimitedFrom(input); 4797 } 4798 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom( 4799 java.io.InputStream input, 4800 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4801 throws java.io.IOException { 4802 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4803 } 4804 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4805 com.google.protobuf.CodedInputStream input) 4806 throws java.io.IOException { 4807 return PARSER.parseFrom(input); 4808 } 4809 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom( 4810 com.google.protobuf.CodedInputStream input, 4811 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4812 throws java.io.IOException { 4813 return PARSER.parseFrom(input, extensionRegistry); 4814 } 4815 4816 public static Builder newBuilder() { return Builder.create(); } 4817 public Builder newBuilderForType() { return newBuilder(); } 4818 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto prototype) { 4819 return newBuilder().mergeFrom(prototype); 4820 } 4821 public Builder toBuilder() { return newBuilder(this); } 4822 4823 @java.lang.Override 4824 protected Builder newBuilderForType( 4825 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4826 Builder builder = new Builder(parent); 4827 return builder; 4828 } 4829 /** 4830 * Protobuf type {@code hadoop.hdfs.HeartbeatResponseProto} 4831 * 4832 * <pre> 4833 * void response 4834 * </pre> 4835 */ 4836 public static final class Builder extends 4837 com.google.protobuf.GeneratedMessage.Builder<Builder> 4838 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProtoOrBuilder { 4839 public static final com.google.protobuf.Descriptors.Descriptor 4840 getDescriptor() { 4841 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor; 4842 } 4843 4844 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4845 internalGetFieldAccessorTable() { 4846 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable 4847 .ensureFieldAccessorsInitialized( 4848 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class); 4849 } 4850 4851 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.newBuilder() 4852 private Builder() { 4853 maybeForceBuilderInitialization(); 4854 } 4855 4856 private Builder( 4857 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4858 super(parent); 4859 maybeForceBuilderInitialization(); 4860 } 4861 private void maybeForceBuilderInitialization() { 4862 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4863 } 4864 } 4865 private static Builder create() { 4866 return new Builder(); 4867 } 4868 4869 public Builder clear() { 4870 super.clear(); 4871 return this; 4872 } 4873 4874 public Builder clone() { 4875 return create().mergeFrom(buildPartial()); 4876 } 4877 4878 public com.google.protobuf.Descriptors.Descriptor 4879 getDescriptorForType() { 4880 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor; 4881 } 4882 4883 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() { 4884 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); 4885 } 4886 4887 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto build() { 4888 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = buildPartial(); 4889 if (!result.isInitialized()) { 4890 throw newUninitializedMessageException(result); 4891 } 4892 return result; 4893 } 4894 4895 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto buildPartial() { 4896 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto(this); 4897 onBuilt(); 4898 return result; 4899 } 4900 4901 public Builder mergeFrom(com.google.protobuf.Message other) { 4902 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) { 4903 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)other); 4904 } else { 4905 super.mergeFrom(other); 4906 return this; 4907 } 4908 } 4909 4910 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other) { 4911 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()) return this; 4912 this.mergeUnknownFields(other.getUnknownFields()); 4913 return this; 4914 } 4915 4916 public final boolean isInitialized() { 4917 return true; 4918 } 4919 4920 public Builder mergeFrom( 4921 com.google.protobuf.CodedInputStream input, 4922 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4923 throws java.io.IOException { 4924 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parsedMessage = null; 4925 try { 4926 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4927 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4928 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) e.getUnfinishedMessage(); 4929 throw e; 4930 } finally { 4931 if (parsedMessage != null) { 4932 mergeFrom(parsedMessage); 4933 } 4934 } 4935 return this; 4936 } 4937 4938 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatResponseProto) 4939 } 4940 4941 static { 4942 defaultInstance = new HeartbeatResponseProto(true); 4943 defaultInstance.initFields(); 4944 } 4945 4946 // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatResponseProto) 4947 } 4948 4949 public interface StartLogSegmentRequestProtoOrBuilder 4950 extends com.google.protobuf.MessageOrBuilder { 4951 4952 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 4953 /** 4954 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4955 */ 4956 boolean hasReqInfo(); 4957 /** 4958 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4959 */ 4960 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 4961 /** 4962 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 4963 */ 4964 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 4965 4966 // required uint64 txid = 2; 4967 /** 4968 * <code>required uint64 txid = 2;</code> 4969 * 4970 * <pre> 4971 * Transaction ID 4972 * </pre> 4973 */ 4974 boolean hasTxid(); 4975 /** 4976 * <code>required uint64 txid = 2;</code> 4977 * 4978 * <pre> 4979 * Transaction ID 4980 * </pre> 4981 */ 4982 long getTxid(); 4983 } 4984 /** 4985 * Protobuf type {@code hadoop.hdfs.StartLogSegmentRequestProto} 4986 * 4987 * <pre> 4988 ** 4989 * startLogSegment() 4990 * </pre> 4991 */ 4992 public static final class StartLogSegmentRequestProto extends 4993 com.google.protobuf.GeneratedMessage 4994 implements StartLogSegmentRequestProtoOrBuilder { 4995 // Use StartLogSegmentRequestProto.newBuilder() to construct. 4996 private StartLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4997 super(builder); 4998 this.unknownFields = builder.getUnknownFields(); 4999 } 5000 private StartLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 5001 5002 private static final StartLogSegmentRequestProto defaultInstance; 5003 public static StartLogSegmentRequestProto getDefaultInstance() { 5004 return defaultInstance; 5005 } 5006 5007 public StartLogSegmentRequestProto getDefaultInstanceForType() { 5008 return defaultInstance; 5009 } 5010 5011 private final com.google.protobuf.UnknownFieldSet unknownFields; 5012 @java.lang.Override 5013 public final com.google.protobuf.UnknownFieldSet 5014 getUnknownFields() { 5015 return this.unknownFields; 5016 } 5017 private StartLogSegmentRequestProto( 5018 com.google.protobuf.CodedInputStream input, 5019 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5020 throws com.google.protobuf.InvalidProtocolBufferException { 5021 initFields(); 5022 int mutable_bitField0_ = 0; 5023 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 5024 com.google.protobuf.UnknownFieldSet.newBuilder(); 5025 try { 5026 boolean done = false; 5027 while (!done) { 5028 int tag = input.readTag(); 5029 switch (tag) { 5030 case 0: 5031 done = true; 5032 break; 5033 default: { 5034 if (!parseUnknownField(input, unknownFields, 5035 extensionRegistry, tag)) { 5036 done = true; 5037 } 5038 break; 5039 } 5040 case 10: { 5041 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 5042 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5043 subBuilder = reqInfo_.toBuilder(); 5044 } 5045 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 5046 if (subBuilder != null) { 5047 subBuilder.mergeFrom(reqInfo_); 5048 reqInfo_ = subBuilder.buildPartial(); 5049 } 5050 bitField0_ |= 0x00000001; 5051 break; 5052 } 5053 case 16: { 5054 bitField0_ |= 0x00000002; 5055 txid_ = input.readUInt64(); 5056 break; 5057 } 5058 } 5059 } 5060 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5061 throw e.setUnfinishedMessage(this); 5062 } catch (java.io.IOException e) { 5063 throw new com.google.protobuf.InvalidProtocolBufferException( 5064 e.getMessage()).setUnfinishedMessage(this); 5065 } finally { 5066 this.unknownFields = unknownFields.build(); 5067 makeExtensionsImmutable(); 5068 } 5069 } 5070 public static final com.google.protobuf.Descriptors.Descriptor 5071 getDescriptor() { 5072 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor; 5073 } 5074 5075 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5076 internalGetFieldAccessorTable() { 5077 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable 5078 .ensureFieldAccessorsInitialized( 5079 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class); 5080 } 5081 5082 public static com.google.protobuf.Parser<StartLogSegmentRequestProto> PARSER = 5083 new com.google.protobuf.AbstractParser<StartLogSegmentRequestProto>() { 5084 public StartLogSegmentRequestProto parsePartialFrom( 5085 com.google.protobuf.CodedInputStream input, 5086 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5087 throws com.google.protobuf.InvalidProtocolBufferException { 5088 return new StartLogSegmentRequestProto(input, extensionRegistry); 5089 } 5090 }; 5091 5092 @java.lang.Override 5093 public com.google.protobuf.Parser<StartLogSegmentRequestProto> getParserForType() { 5094 return PARSER; 5095 } 5096 5097 private int bitField0_; 5098 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 5099 public static final int REQINFO_FIELD_NUMBER = 1; 5100 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 5101 /** 5102 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5103 */ 5104 public boolean hasReqInfo() { 5105 return ((bitField0_ & 0x00000001) == 0x00000001); 5106 } 5107 /** 5108 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5109 */ 5110 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 5111 return reqInfo_; 5112 } 5113 /** 5114 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5115 */ 5116 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 5117 return reqInfo_; 5118 } 5119 5120 // required uint64 txid = 2; 5121 public static final int TXID_FIELD_NUMBER = 2; 5122 private long txid_; 5123 /** 5124 * <code>required uint64 txid = 2;</code> 5125 * 5126 * <pre> 5127 * Transaction ID 5128 * </pre> 5129 */ 5130 public boolean hasTxid() { 5131 return ((bitField0_ & 0x00000002) == 0x00000002); 5132 } 5133 /** 5134 * <code>required uint64 txid = 2;</code> 5135 * 5136 * <pre> 5137 * Transaction ID 5138 * </pre> 5139 */ 5140 public long getTxid() { 5141 return txid_; 5142 } 5143 5144 private void initFields() { 5145 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 5146 txid_ = 0L; 5147 } 5148 private byte memoizedIsInitialized = -1; 5149 public final boolean isInitialized() { 5150 byte isInitialized = memoizedIsInitialized; 5151 if (isInitialized != -1) return isInitialized == 1; 5152 5153 if (!hasReqInfo()) { 5154 memoizedIsInitialized = 0; 5155 return false; 5156 } 5157 if (!hasTxid()) { 5158 memoizedIsInitialized = 0; 5159 return false; 5160 } 5161 if (!getReqInfo().isInitialized()) { 5162 memoizedIsInitialized = 0; 5163 return false; 5164 } 5165 memoizedIsInitialized = 1; 5166 return true; 5167 } 5168 5169 public void writeTo(com.google.protobuf.CodedOutputStream output) 5170 throws java.io.IOException { 5171 getSerializedSize(); 5172 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5173 output.writeMessage(1, reqInfo_); 5174 } 5175 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5176 output.writeUInt64(2, txid_); 5177 } 5178 getUnknownFields().writeTo(output); 5179 } 5180 5181 private int memoizedSerializedSize = -1; 5182 public int getSerializedSize() { 5183 int size = memoizedSerializedSize; 5184 if (size != -1) return size; 5185 5186 size = 0; 5187 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5188 size += com.google.protobuf.CodedOutputStream 5189 .computeMessageSize(1, reqInfo_); 5190 } 5191 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5192 size += com.google.protobuf.CodedOutputStream 5193 .computeUInt64Size(2, txid_); 5194 } 5195 size += getUnknownFields().getSerializedSize(); 5196 memoizedSerializedSize = size; 5197 return size; 5198 } 5199 5200 private static final long serialVersionUID = 0L; 5201 @java.lang.Override 5202 protected java.lang.Object writeReplace() 5203 throws java.io.ObjectStreamException { 5204 return super.writeReplace(); 5205 } 5206 5207 @java.lang.Override 5208 public boolean equals(final java.lang.Object obj) { 5209 if (obj == this) { 5210 return true; 5211 } 5212 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)) { 5213 return super.equals(obj); 5214 } 5215 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) obj; 5216 5217 boolean result = true; 5218 result = result && (hasReqInfo() == other.hasReqInfo()); 5219 if (hasReqInfo()) { 5220 result = result && getReqInfo() 5221 .equals(other.getReqInfo()); 5222 } 5223 result = result && (hasTxid() == other.hasTxid()); 5224 if (hasTxid()) { 5225 result = result && (getTxid() 5226 == other.getTxid()); 5227 } 5228 result = result && 5229 getUnknownFields().equals(other.getUnknownFields()); 5230 return result; 5231 } 5232 5233 private int memoizedHashCode = 0; 5234 @java.lang.Override 5235 public int hashCode() { 5236 if (memoizedHashCode != 0) { 5237 return memoizedHashCode; 5238 } 5239 int hash = 41; 5240 hash = (19 * hash) + getDescriptorForType().hashCode(); 5241 if (hasReqInfo()) { 5242 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 5243 hash = (53 * hash) + getReqInfo().hashCode(); 5244 } 5245 if (hasTxid()) { 5246 hash = (37 * hash) + TXID_FIELD_NUMBER; 5247 hash = (53 * hash) + hashLong(getTxid()); 5248 } 5249 hash = (29 * hash) + getUnknownFields().hashCode(); 5250 memoizedHashCode = hash; 5251 return hash; 5252 } 5253 5254 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5255 com.google.protobuf.ByteString data) 5256 throws com.google.protobuf.InvalidProtocolBufferException { 5257 return PARSER.parseFrom(data); 5258 } 5259 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5260 com.google.protobuf.ByteString data, 5261 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5262 throws com.google.protobuf.InvalidProtocolBufferException { 5263 return PARSER.parseFrom(data, extensionRegistry); 5264 } 5265 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(byte[] data) 5266 throws com.google.protobuf.InvalidProtocolBufferException { 5267 return PARSER.parseFrom(data); 5268 } 5269 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5270 byte[] data, 5271 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5272 throws com.google.protobuf.InvalidProtocolBufferException { 5273 return PARSER.parseFrom(data, extensionRegistry); 5274 } 5275 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(java.io.InputStream input) 5276 throws java.io.IOException { 5277 return PARSER.parseFrom(input); 5278 } 5279 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5280 java.io.InputStream input, 5281 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5282 throws java.io.IOException { 5283 return PARSER.parseFrom(input, extensionRegistry); 5284 } 5285 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input) 5286 throws java.io.IOException { 5287 return PARSER.parseDelimitedFrom(input); 5288 } 5289 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom( 5290 java.io.InputStream input, 5291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5292 throws java.io.IOException { 5293 return PARSER.parseDelimitedFrom(input, extensionRegistry); 5294 } 5295 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5296 com.google.protobuf.CodedInputStream input) 5297 throws java.io.IOException { 5298 return PARSER.parseFrom(input); 5299 } 5300 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom( 5301 com.google.protobuf.CodedInputStream input, 5302 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5303 throws java.io.IOException { 5304 return PARSER.parseFrom(input, extensionRegistry); 5305 } 5306 5307 public static Builder newBuilder() { return Builder.create(); } 5308 public Builder newBuilderForType() { return newBuilder(); } 5309 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto prototype) { 5310 return newBuilder().mergeFrom(prototype); 5311 } 5312 public Builder toBuilder() { return newBuilder(this); } 5313 5314 @java.lang.Override 5315 protected Builder newBuilderForType( 5316 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5317 Builder builder = new Builder(parent); 5318 return builder; 5319 } 5320 /** 5321 * Protobuf type {@code hadoop.hdfs.StartLogSegmentRequestProto} 5322 * 5323 * <pre> 5324 ** 5325 * startLogSegment() 5326 * </pre> 5327 */ 5328 public static final class Builder extends 5329 com.google.protobuf.GeneratedMessage.Builder<Builder> 5330 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProtoOrBuilder { 5331 public static final com.google.protobuf.Descriptors.Descriptor 5332 getDescriptor() { 5333 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor; 5334 } 5335 5336 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5337 internalGetFieldAccessorTable() { 5338 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable 5339 .ensureFieldAccessorsInitialized( 5340 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class); 5341 } 5342 5343 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.newBuilder() 5344 private Builder() { 5345 maybeForceBuilderInitialization(); 5346 } 5347 5348 private Builder( 5349 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5350 super(parent); 5351 maybeForceBuilderInitialization(); 5352 } 5353 private void maybeForceBuilderInitialization() { 5354 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 5355 getReqInfoFieldBuilder(); 5356 } 5357 } 5358 private static Builder create() { 5359 return new Builder(); 5360 } 5361 5362 public Builder clear() { 5363 super.clear(); 5364 if (reqInfoBuilder_ == null) { 5365 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 5366 } else { 5367 reqInfoBuilder_.clear(); 5368 } 5369 bitField0_ = (bitField0_ & ~0x00000001); 5370 txid_ = 0L; 5371 bitField0_ = (bitField0_ & ~0x00000002); 5372 return this; 5373 } 5374 5375 public Builder clone() { 5376 return create().mergeFrom(buildPartial()); 5377 } 5378 5379 public com.google.protobuf.Descriptors.Descriptor 5380 getDescriptorForType() { 5381 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor; 5382 } 5383 5384 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto getDefaultInstanceForType() { 5385 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance(); 5386 } 5387 5388 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto build() { 5389 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = buildPartial(); 5390 if (!result.isInitialized()) { 5391 throw newUninitializedMessageException(result); 5392 } 5393 return result; 5394 } 5395 5396 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto buildPartial() { 5397 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto(this); 5398 int from_bitField0_ = bitField0_; 5399 int to_bitField0_ = 0; 5400 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 5401 to_bitField0_ |= 0x00000001; 5402 } 5403 if (reqInfoBuilder_ == null) { 5404 result.reqInfo_ = reqInfo_; 5405 } else { 5406 result.reqInfo_ = reqInfoBuilder_.build(); 5407 } 5408 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 5409 to_bitField0_ |= 0x00000002; 5410 } 5411 result.txid_ = txid_; 5412 result.bitField0_ = to_bitField0_; 5413 onBuilt(); 5414 return result; 5415 } 5416 5417 public Builder mergeFrom(com.google.protobuf.Message other) { 5418 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) { 5419 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)other); 5420 } else { 5421 super.mergeFrom(other); 5422 return this; 5423 } 5424 } 5425 5426 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other) { 5427 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance()) return this; 5428 if (other.hasReqInfo()) { 5429 mergeReqInfo(other.getReqInfo()); 5430 } 5431 if (other.hasTxid()) { 5432 setTxid(other.getTxid()); 5433 } 5434 this.mergeUnknownFields(other.getUnknownFields()); 5435 return this; 5436 } 5437 5438 public final boolean isInitialized() { 5439 if (!hasReqInfo()) { 5440 5441 return false; 5442 } 5443 if (!hasTxid()) { 5444 5445 return false; 5446 } 5447 if (!getReqInfo().isInitialized()) { 5448 5449 return false; 5450 } 5451 return true; 5452 } 5453 5454 public Builder mergeFrom( 5455 com.google.protobuf.CodedInputStream input, 5456 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5457 throws java.io.IOException { 5458 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parsedMessage = null; 5459 try { 5460 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 5461 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5462 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) e.getUnfinishedMessage(); 5463 throw e; 5464 } finally { 5465 if (parsedMessage != null) { 5466 mergeFrom(parsedMessage); 5467 } 5468 } 5469 return this; 5470 } 5471 private int bitField0_; 5472 5473 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 5474 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 5475 private com.google.protobuf.SingleFieldBuilder< 5476 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 5477 /** 5478 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5479 */ 5480 public boolean hasReqInfo() { 5481 return ((bitField0_ & 0x00000001) == 0x00000001); 5482 } 5483 /** 5484 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5485 */ 5486 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 5487 if (reqInfoBuilder_ == null) { 5488 return reqInfo_; 5489 } else { 5490 return reqInfoBuilder_.getMessage(); 5491 } 5492 } 5493 /** 5494 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5495 */ 5496 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 5497 if (reqInfoBuilder_ == null) { 5498 if (value == null) { 5499 throw new NullPointerException(); 5500 } 5501 reqInfo_ = value; 5502 onChanged(); 5503 } else { 5504 reqInfoBuilder_.setMessage(value); 5505 } 5506 bitField0_ |= 0x00000001; 5507 return this; 5508 } 5509 /** 5510 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5511 */ 5512 public Builder setReqInfo( 5513 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 5514 if (reqInfoBuilder_ == null) { 5515 reqInfo_ = builderForValue.build(); 5516 onChanged(); 5517 } else { 5518 reqInfoBuilder_.setMessage(builderForValue.build()); 5519 } 5520 bitField0_ |= 0x00000001; 5521 return this; 5522 } 5523 /** 5524 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5525 */ 5526 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 5527 if (reqInfoBuilder_ == null) { 5528 if (((bitField0_ & 0x00000001) == 0x00000001) && 5529 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 5530 reqInfo_ = 5531 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 5532 } else { 5533 reqInfo_ = value; 5534 } 5535 onChanged(); 5536 } else { 5537 reqInfoBuilder_.mergeFrom(value); 5538 } 5539 bitField0_ |= 0x00000001; 5540 return this; 5541 } 5542 /** 5543 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5544 */ 5545 public Builder clearReqInfo() { 5546 if (reqInfoBuilder_ == null) { 5547 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 5548 onChanged(); 5549 } else { 5550 reqInfoBuilder_.clear(); 5551 } 5552 bitField0_ = (bitField0_ & ~0x00000001); 5553 return this; 5554 } 5555 /** 5556 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5557 */ 5558 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 5559 bitField0_ |= 0x00000001; 5560 onChanged(); 5561 return getReqInfoFieldBuilder().getBuilder(); 5562 } 5563 /** 5564 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5565 */ 5566 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 5567 if (reqInfoBuilder_ != null) { 5568 return reqInfoBuilder_.getMessageOrBuilder(); 5569 } else { 5570 return reqInfo_; 5571 } 5572 } 5573 /** 5574 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5575 */ 5576 private com.google.protobuf.SingleFieldBuilder< 5577 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 5578 getReqInfoFieldBuilder() { 5579 if (reqInfoBuilder_ == null) { 5580 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 5581 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 5582 reqInfo_, 5583 getParentForChildren(), 5584 isClean()); 5585 reqInfo_ = null; 5586 } 5587 return reqInfoBuilder_; 5588 } 5589 5590 // required uint64 txid = 2; 5591 private long txid_ ; 5592 /** 5593 * <code>required uint64 txid = 2;</code> 5594 * 5595 * <pre> 5596 * Transaction ID 5597 * </pre> 5598 */ 5599 public boolean hasTxid() { 5600 return ((bitField0_ & 0x00000002) == 0x00000002); 5601 } 5602 /** 5603 * <code>required uint64 txid = 2;</code> 5604 * 5605 * <pre> 5606 * Transaction ID 5607 * </pre> 5608 */ 5609 public long getTxid() { 5610 return txid_; 5611 } 5612 /** 5613 * <code>required uint64 txid = 2;</code> 5614 * 5615 * <pre> 5616 * Transaction ID 5617 * </pre> 5618 */ 5619 public Builder setTxid(long value) { 5620 bitField0_ |= 0x00000002; 5621 txid_ = value; 5622 onChanged(); 5623 return this; 5624 } 5625 /** 5626 * <code>required uint64 txid = 2;</code> 5627 * 5628 * <pre> 5629 * Transaction ID 5630 * </pre> 5631 */ 5632 public Builder clearTxid() { 5633 bitField0_ = (bitField0_ & ~0x00000002); 5634 txid_ = 0L; 5635 onChanged(); 5636 return this; 5637 } 5638 5639 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentRequestProto) 5640 } 5641 5642 static { 5643 defaultInstance = new StartLogSegmentRequestProto(true); 5644 defaultInstance.initFields(); 5645 } 5646 5647 // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentRequestProto) 5648 } 5649 5650 public interface StartLogSegmentResponseProtoOrBuilder 5651 extends com.google.protobuf.MessageOrBuilder { 5652 } 5653 /** 5654 * Protobuf type {@code hadoop.hdfs.StartLogSegmentResponseProto} 5655 */ 5656 public static final class StartLogSegmentResponseProto extends 5657 com.google.protobuf.GeneratedMessage 5658 implements StartLogSegmentResponseProtoOrBuilder { 5659 // Use StartLogSegmentResponseProto.newBuilder() to construct. 5660 private StartLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 5661 super(builder); 5662 this.unknownFields = builder.getUnknownFields(); 5663 } 5664 private StartLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 5665 5666 private static final StartLogSegmentResponseProto defaultInstance; 5667 public static StartLogSegmentResponseProto getDefaultInstance() { 5668 return defaultInstance; 5669 } 5670 5671 public StartLogSegmentResponseProto getDefaultInstanceForType() { 5672 return defaultInstance; 5673 } 5674 5675 private final com.google.protobuf.UnknownFieldSet unknownFields; 5676 @java.lang.Override 5677 public final com.google.protobuf.UnknownFieldSet 5678 getUnknownFields() { 5679 return this.unknownFields; 5680 } 5681 private StartLogSegmentResponseProto( 5682 com.google.protobuf.CodedInputStream input, 5683 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5684 throws com.google.protobuf.InvalidProtocolBufferException { 5685 initFields(); 5686 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 5687 com.google.protobuf.UnknownFieldSet.newBuilder(); 5688 try { 5689 boolean done = false; 5690 while (!done) { 5691 int tag = input.readTag(); 5692 switch (tag) { 5693 case 0: 5694 done = true; 5695 break; 5696 default: { 5697 if (!parseUnknownField(input, unknownFields, 5698 extensionRegistry, tag)) { 5699 done = true; 5700 } 5701 break; 5702 } 5703 } 5704 } 5705 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5706 throw e.setUnfinishedMessage(this); 5707 } catch (java.io.IOException e) { 5708 throw new com.google.protobuf.InvalidProtocolBufferException( 5709 e.getMessage()).setUnfinishedMessage(this); 5710 } finally { 5711 this.unknownFields = unknownFields.build(); 5712 makeExtensionsImmutable(); 5713 } 5714 } 5715 public static final com.google.protobuf.Descriptors.Descriptor 5716 getDescriptor() { 5717 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor; 5718 } 5719 5720 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5721 internalGetFieldAccessorTable() { 5722 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable 5723 .ensureFieldAccessorsInitialized( 5724 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class); 5725 } 5726 5727 public static com.google.protobuf.Parser<StartLogSegmentResponseProto> PARSER = 5728 new com.google.protobuf.AbstractParser<StartLogSegmentResponseProto>() { 5729 public StartLogSegmentResponseProto parsePartialFrom( 5730 com.google.protobuf.CodedInputStream input, 5731 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5732 throws com.google.protobuf.InvalidProtocolBufferException { 5733 return new StartLogSegmentResponseProto(input, extensionRegistry); 5734 } 5735 }; 5736 5737 @java.lang.Override 5738 public com.google.protobuf.Parser<StartLogSegmentResponseProto> getParserForType() { 5739 return PARSER; 5740 } 5741 5742 private void initFields() { 5743 } 5744 private byte memoizedIsInitialized = -1; 5745 public final boolean isInitialized() { 5746 byte isInitialized = memoizedIsInitialized; 5747 if (isInitialized != -1) return isInitialized == 1; 5748 5749 memoizedIsInitialized = 1; 5750 return true; 5751 } 5752 5753 public void writeTo(com.google.protobuf.CodedOutputStream output) 5754 throws java.io.IOException { 5755 getSerializedSize(); 5756 getUnknownFields().writeTo(output); 5757 } 5758 5759 private int memoizedSerializedSize = -1; 5760 public int getSerializedSize() { 5761 int size = memoizedSerializedSize; 5762 if (size != -1) return size; 5763 5764 size = 0; 5765 size += getUnknownFields().getSerializedSize(); 5766 memoizedSerializedSize = size; 5767 return size; 5768 } 5769 5770 private static final long serialVersionUID = 0L; 5771 @java.lang.Override 5772 protected java.lang.Object writeReplace() 5773 throws java.io.ObjectStreamException { 5774 return super.writeReplace(); 5775 } 5776 5777 @java.lang.Override 5778 public boolean equals(final java.lang.Object obj) { 5779 if (obj == this) { 5780 return true; 5781 } 5782 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)) { 5783 return super.equals(obj); 5784 } 5785 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) obj; 5786 5787 boolean result = true; 5788 result = result && 5789 getUnknownFields().equals(other.getUnknownFields()); 5790 return result; 5791 } 5792 5793 private int memoizedHashCode = 0; 5794 @java.lang.Override 5795 public int hashCode() { 5796 if (memoizedHashCode != 0) { 5797 return memoizedHashCode; 5798 } 5799 int hash = 41; 5800 hash = (19 * hash) + getDescriptorForType().hashCode(); 5801 hash = (29 * hash) + getUnknownFields().hashCode(); 5802 memoizedHashCode = hash; 5803 return hash; 5804 } 5805 5806 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5807 com.google.protobuf.ByteString data) 5808 throws com.google.protobuf.InvalidProtocolBufferException { 5809 return PARSER.parseFrom(data); 5810 } 5811 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5812 com.google.protobuf.ByteString data, 5813 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5814 throws com.google.protobuf.InvalidProtocolBufferException { 5815 return PARSER.parseFrom(data, extensionRegistry); 5816 } 5817 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(byte[] data) 5818 throws com.google.protobuf.InvalidProtocolBufferException { 5819 return PARSER.parseFrom(data); 5820 } 5821 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5822 byte[] data, 5823 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5824 throws com.google.protobuf.InvalidProtocolBufferException { 5825 return PARSER.parseFrom(data, extensionRegistry); 5826 } 5827 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(java.io.InputStream input) 5828 throws java.io.IOException { 5829 return PARSER.parseFrom(input); 5830 } 5831 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5832 java.io.InputStream input, 5833 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5834 throws java.io.IOException { 5835 return PARSER.parseFrom(input, extensionRegistry); 5836 } 5837 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input) 5838 throws java.io.IOException { 5839 return PARSER.parseDelimitedFrom(input); 5840 } 5841 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom( 5842 java.io.InputStream input, 5843 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5844 throws java.io.IOException { 5845 return PARSER.parseDelimitedFrom(input, extensionRegistry); 5846 } 5847 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5848 com.google.protobuf.CodedInputStream input) 5849 throws java.io.IOException { 5850 return PARSER.parseFrom(input); 5851 } 5852 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom( 5853 com.google.protobuf.CodedInputStream input, 5854 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5855 throws java.io.IOException { 5856 return PARSER.parseFrom(input, extensionRegistry); 5857 } 5858 5859 public static Builder newBuilder() { return Builder.create(); } 5860 public Builder newBuilderForType() { return newBuilder(); } 5861 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto prototype) { 5862 return newBuilder().mergeFrom(prototype); 5863 } 5864 public Builder toBuilder() { return newBuilder(this); } 5865 5866 @java.lang.Override 5867 protected Builder newBuilderForType( 5868 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5869 Builder builder = new Builder(parent); 5870 return builder; 5871 } 5872 /** 5873 * Protobuf type {@code hadoop.hdfs.StartLogSegmentResponseProto} 5874 */ 5875 public static final class Builder extends 5876 com.google.protobuf.GeneratedMessage.Builder<Builder> 5877 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProtoOrBuilder { 5878 public static final com.google.protobuf.Descriptors.Descriptor 5879 getDescriptor() { 5880 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor; 5881 } 5882 5883 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5884 internalGetFieldAccessorTable() { 5885 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable 5886 .ensureFieldAccessorsInitialized( 5887 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class); 5888 } 5889 5890 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.newBuilder() 5891 private Builder() { 5892 maybeForceBuilderInitialization(); 5893 } 5894 5895 private Builder( 5896 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5897 super(parent); 5898 maybeForceBuilderInitialization(); 5899 } 5900 private void maybeForceBuilderInitialization() { 5901 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 5902 } 5903 } 5904 private static Builder create() { 5905 return new Builder(); 5906 } 5907 5908 public Builder clear() { 5909 super.clear(); 5910 return this; 5911 } 5912 5913 public Builder clone() { 5914 return create().mergeFrom(buildPartial()); 5915 } 5916 5917 public com.google.protobuf.Descriptors.Descriptor 5918 getDescriptorForType() { 5919 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor; 5920 } 5921 5922 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto getDefaultInstanceForType() { 5923 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(); 5924 } 5925 5926 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto build() { 5927 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = buildPartial(); 5928 if (!result.isInitialized()) { 5929 throw newUninitializedMessageException(result); 5930 } 5931 return result; 5932 } 5933 5934 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto buildPartial() { 5935 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto(this); 5936 onBuilt(); 5937 return result; 5938 } 5939 5940 public Builder mergeFrom(com.google.protobuf.Message other) { 5941 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) { 5942 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)other); 5943 } else { 5944 super.mergeFrom(other); 5945 return this; 5946 } 5947 } 5948 5949 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other) { 5950 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()) return this; 5951 this.mergeUnknownFields(other.getUnknownFields()); 5952 return this; 5953 } 5954 5955 public final boolean isInitialized() { 5956 return true; 5957 } 5958 5959 public Builder mergeFrom( 5960 com.google.protobuf.CodedInputStream input, 5961 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5962 throws java.io.IOException { 5963 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parsedMessage = null; 5964 try { 5965 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 5966 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5967 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) e.getUnfinishedMessage(); 5968 throw e; 5969 } finally { 5970 if (parsedMessage != null) { 5971 mergeFrom(parsedMessage); 5972 } 5973 } 5974 return this; 5975 } 5976 5977 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentResponseProto) 5978 } 5979 5980 static { 5981 defaultInstance = new StartLogSegmentResponseProto(true); 5982 defaultInstance.initFields(); 5983 } 5984 5985 // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentResponseProto) 5986 } 5987 5988 public interface FinalizeLogSegmentRequestProtoOrBuilder 5989 extends com.google.protobuf.MessageOrBuilder { 5990 5991 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 5992 /** 5993 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5994 */ 5995 boolean hasReqInfo(); 5996 /** 5997 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 5998 */ 5999 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 6000 /** 6001 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6002 */ 6003 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 6004 6005 // required uint64 startTxId = 2; 6006 /** 6007 * <code>required uint64 startTxId = 2;</code> 6008 */ 6009 boolean hasStartTxId(); 6010 /** 6011 * <code>required uint64 startTxId = 2;</code> 6012 */ 6013 long getStartTxId(); 6014 6015 // required uint64 endTxId = 3; 6016 /** 6017 * <code>required uint64 endTxId = 3;</code> 6018 */ 6019 boolean hasEndTxId(); 6020 /** 6021 * <code>required uint64 endTxId = 3;</code> 6022 */ 6023 long getEndTxId(); 6024 } 6025 /** 6026 * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentRequestProto} 6027 * 6028 * <pre> 6029 ** 6030 * finalizeLogSegment() 6031 * </pre> 6032 */ 6033 public static final class FinalizeLogSegmentRequestProto extends 6034 com.google.protobuf.GeneratedMessage 6035 implements FinalizeLogSegmentRequestProtoOrBuilder { 6036 // Use FinalizeLogSegmentRequestProto.newBuilder() to construct. 6037 private FinalizeLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 6038 super(builder); 6039 this.unknownFields = builder.getUnknownFields(); 6040 } 6041 private FinalizeLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 6042 6043 private static final FinalizeLogSegmentRequestProto defaultInstance; 6044 public static FinalizeLogSegmentRequestProto getDefaultInstance() { 6045 return defaultInstance; 6046 } 6047 6048 public FinalizeLogSegmentRequestProto getDefaultInstanceForType() { 6049 return defaultInstance; 6050 } 6051 6052 private final com.google.protobuf.UnknownFieldSet unknownFields; 6053 @java.lang.Override 6054 public final com.google.protobuf.UnknownFieldSet 6055 getUnknownFields() { 6056 return this.unknownFields; 6057 } 6058 private FinalizeLogSegmentRequestProto( 6059 com.google.protobuf.CodedInputStream input, 6060 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6061 throws com.google.protobuf.InvalidProtocolBufferException { 6062 initFields(); 6063 int mutable_bitField0_ = 0; 6064 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 6065 com.google.protobuf.UnknownFieldSet.newBuilder(); 6066 try { 6067 boolean done = false; 6068 while (!done) { 6069 int tag = input.readTag(); 6070 switch (tag) { 6071 case 0: 6072 done = true; 6073 break; 6074 default: { 6075 if (!parseUnknownField(input, unknownFields, 6076 extensionRegistry, tag)) { 6077 done = true; 6078 } 6079 break; 6080 } 6081 case 10: { 6082 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 6083 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6084 subBuilder = reqInfo_.toBuilder(); 6085 } 6086 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 6087 if (subBuilder != null) { 6088 subBuilder.mergeFrom(reqInfo_); 6089 reqInfo_ = subBuilder.buildPartial(); 6090 } 6091 bitField0_ |= 0x00000001; 6092 break; 6093 } 6094 case 16: { 6095 bitField0_ |= 0x00000002; 6096 startTxId_ = input.readUInt64(); 6097 break; 6098 } 6099 case 24: { 6100 bitField0_ |= 0x00000004; 6101 endTxId_ = input.readUInt64(); 6102 break; 6103 } 6104 } 6105 } 6106 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6107 throw e.setUnfinishedMessage(this); 6108 } catch (java.io.IOException e) { 6109 throw new com.google.protobuf.InvalidProtocolBufferException( 6110 e.getMessage()).setUnfinishedMessage(this); 6111 } finally { 6112 this.unknownFields = unknownFields.build(); 6113 makeExtensionsImmutable(); 6114 } 6115 } 6116 public static final com.google.protobuf.Descriptors.Descriptor 6117 getDescriptor() { 6118 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor; 6119 } 6120 6121 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6122 internalGetFieldAccessorTable() { 6123 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable 6124 .ensureFieldAccessorsInitialized( 6125 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class); 6126 } 6127 6128 public static com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> PARSER = 6129 new com.google.protobuf.AbstractParser<FinalizeLogSegmentRequestProto>() { 6130 public FinalizeLogSegmentRequestProto parsePartialFrom( 6131 com.google.protobuf.CodedInputStream input, 6132 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6133 throws com.google.protobuf.InvalidProtocolBufferException { 6134 return new FinalizeLogSegmentRequestProto(input, extensionRegistry); 6135 } 6136 }; 6137 6138 @java.lang.Override 6139 public com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> getParserForType() { 6140 return PARSER; 6141 } 6142 6143 private int bitField0_; 6144 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 6145 public static final int REQINFO_FIELD_NUMBER = 1; 6146 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 6147 /** 6148 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6149 */ 6150 public boolean hasReqInfo() { 6151 return ((bitField0_ & 0x00000001) == 0x00000001); 6152 } 6153 /** 6154 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6155 */ 6156 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 6157 return reqInfo_; 6158 } 6159 /** 6160 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6161 */ 6162 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 6163 return reqInfo_; 6164 } 6165 6166 // required uint64 startTxId = 2; 6167 public static final int STARTTXID_FIELD_NUMBER = 2; 6168 private long startTxId_; 6169 /** 6170 * <code>required uint64 startTxId = 2;</code> 6171 */ 6172 public boolean hasStartTxId() { 6173 return ((bitField0_ & 0x00000002) == 0x00000002); 6174 } 6175 /** 6176 * <code>required uint64 startTxId = 2;</code> 6177 */ 6178 public long getStartTxId() { 6179 return startTxId_; 6180 } 6181 6182 // required uint64 endTxId = 3; 6183 public static final int ENDTXID_FIELD_NUMBER = 3; 6184 private long endTxId_; 6185 /** 6186 * <code>required uint64 endTxId = 3;</code> 6187 */ 6188 public boolean hasEndTxId() { 6189 return ((bitField0_ & 0x00000004) == 0x00000004); 6190 } 6191 /** 6192 * <code>required uint64 endTxId = 3;</code> 6193 */ 6194 public long getEndTxId() { 6195 return endTxId_; 6196 } 6197 6198 private void initFields() { 6199 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 6200 startTxId_ = 0L; 6201 endTxId_ = 0L; 6202 } 6203 private byte memoizedIsInitialized = -1; 6204 public final boolean isInitialized() { 6205 byte isInitialized = memoizedIsInitialized; 6206 if (isInitialized != -1) return isInitialized == 1; 6207 6208 if (!hasReqInfo()) { 6209 memoizedIsInitialized = 0; 6210 return false; 6211 } 6212 if (!hasStartTxId()) { 6213 memoizedIsInitialized = 0; 6214 return false; 6215 } 6216 if (!hasEndTxId()) { 6217 memoizedIsInitialized = 0; 6218 return false; 6219 } 6220 if (!getReqInfo().isInitialized()) { 6221 memoizedIsInitialized = 0; 6222 return false; 6223 } 6224 memoizedIsInitialized = 1; 6225 return true; 6226 } 6227 6228 public void writeTo(com.google.protobuf.CodedOutputStream output) 6229 throws java.io.IOException { 6230 getSerializedSize(); 6231 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6232 output.writeMessage(1, reqInfo_); 6233 } 6234 if (((bitField0_ & 0x00000002) == 0x00000002)) { 6235 output.writeUInt64(2, startTxId_); 6236 } 6237 if (((bitField0_ & 0x00000004) == 0x00000004)) { 6238 output.writeUInt64(3, endTxId_); 6239 } 6240 getUnknownFields().writeTo(output); 6241 } 6242 6243 private int memoizedSerializedSize = -1; 6244 public int getSerializedSize() { 6245 int size = memoizedSerializedSize; 6246 if (size != -1) return size; 6247 6248 size = 0; 6249 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6250 size += com.google.protobuf.CodedOutputStream 6251 .computeMessageSize(1, reqInfo_); 6252 } 6253 if (((bitField0_ & 0x00000002) == 0x00000002)) { 6254 size += com.google.protobuf.CodedOutputStream 6255 .computeUInt64Size(2, startTxId_); 6256 } 6257 if (((bitField0_ & 0x00000004) == 0x00000004)) { 6258 size += com.google.protobuf.CodedOutputStream 6259 .computeUInt64Size(3, endTxId_); 6260 } 6261 size += getUnknownFields().getSerializedSize(); 6262 memoizedSerializedSize = size; 6263 return size; 6264 } 6265 6266 private static final long serialVersionUID = 0L; 6267 @java.lang.Override 6268 protected java.lang.Object writeReplace() 6269 throws java.io.ObjectStreamException { 6270 return super.writeReplace(); 6271 } 6272 6273 @java.lang.Override 6274 public boolean equals(final java.lang.Object obj) { 6275 if (obj == this) { 6276 return true; 6277 } 6278 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)) { 6279 return super.equals(obj); 6280 } 6281 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) obj; 6282 6283 boolean result = true; 6284 result = result && (hasReqInfo() == other.hasReqInfo()); 6285 if (hasReqInfo()) { 6286 result = result && getReqInfo() 6287 .equals(other.getReqInfo()); 6288 } 6289 result = result && (hasStartTxId() == other.hasStartTxId()); 6290 if (hasStartTxId()) { 6291 result = result && (getStartTxId() 6292 == other.getStartTxId()); 6293 } 6294 result = result && (hasEndTxId() == other.hasEndTxId()); 6295 if (hasEndTxId()) { 6296 result = result && (getEndTxId() 6297 == other.getEndTxId()); 6298 } 6299 result = result && 6300 getUnknownFields().equals(other.getUnknownFields()); 6301 return result; 6302 } 6303 6304 private int memoizedHashCode = 0; 6305 @java.lang.Override 6306 public int hashCode() { 6307 if (memoizedHashCode != 0) { 6308 return memoizedHashCode; 6309 } 6310 int hash = 41; 6311 hash = (19 * hash) + getDescriptorForType().hashCode(); 6312 if (hasReqInfo()) { 6313 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 6314 hash = (53 * hash) + getReqInfo().hashCode(); 6315 } 6316 if (hasStartTxId()) { 6317 hash = (37 * hash) + STARTTXID_FIELD_NUMBER; 6318 hash = (53 * hash) + hashLong(getStartTxId()); 6319 } 6320 if (hasEndTxId()) { 6321 hash = (37 * hash) + ENDTXID_FIELD_NUMBER; 6322 hash = (53 * hash) + hashLong(getEndTxId()); 6323 } 6324 hash = (29 * hash) + getUnknownFields().hashCode(); 6325 memoizedHashCode = hash; 6326 return hash; 6327 } 6328 6329 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6330 com.google.protobuf.ByteString data) 6331 throws com.google.protobuf.InvalidProtocolBufferException { 6332 return PARSER.parseFrom(data); 6333 } 6334 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6335 com.google.protobuf.ByteString data, 6336 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6337 throws com.google.protobuf.InvalidProtocolBufferException { 6338 return PARSER.parseFrom(data, extensionRegistry); 6339 } 6340 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(byte[] data) 6341 throws com.google.protobuf.InvalidProtocolBufferException { 6342 return PARSER.parseFrom(data); 6343 } 6344 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6345 byte[] data, 6346 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6347 throws com.google.protobuf.InvalidProtocolBufferException { 6348 return PARSER.parseFrom(data, extensionRegistry); 6349 } 6350 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(java.io.InputStream input) 6351 throws java.io.IOException { 6352 return PARSER.parseFrom(input); 6353 } 6354 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6355 java.io.InputStream input, 6356 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6357 throws java.io.IOException { 6358 return PARSER.parseFrom(input, extensionRegistry); 6359 } 6360 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input) 6361 throws java.io.IOException { 6362 return PARSER.parseDelimitedFrom(input); 6363 } 6364 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom( 6365 java.io.InputStream input, 6366 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6367 throws java.io.IOException { 6368 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6369 } 6370 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6371 com.google.protobuf.CodedInputStream input) 6372 throws java.io.IOException { 6373 return PARSER.parseFrom(input); 6374 } 6375 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom( 6376 com.google.protobuf.CodedInputStream input, 6377 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6378 throws java.io.IOException { 6379 return PARSER.parseFrom(input, extensionRegistry); 6380 } 6381 6382 public static Builder newBuilder() { return Builder.create(); } 6383 public Builder newBuilderForType() { return newBuilder(); } 6384 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto prototype) { 6385 return newBuilder().mergeFrom(prototype); 6386 } 6387 public Builder toBuilder() { return newBuilder(this); } 6388 6389 @java.lang.Override 6390 protected Builder newBuilderForType( 6391 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6392 Builder builder = new Builder(parent); 6393 return builder; 6394 } 6395 /** 6396 * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentRequestProto} 6397 * 6398 * <pre> 6399 ** 6400 * finalizeLogSegment() 6401 * </pre> 6402 */ 6403 public static final class Builder extends 6404 com.google.protobuf.GeneratedMessage.Builder<Builder> 6405 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProtoOrBuilder { 6406 public static final com.google.protobuf.Descriptors.Descriptor 6407 getDescriptor() { 6408 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor; 6409 } 6410 6411 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6412 internalGetFieldAccessorTable() { 6413 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable 6414 .ensureFieldAccessorsInitialized( 6415 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class); 6416 } 6417 6418 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.newBuilder() 6419 private Builder() { 6420 maybeForceBuilderInitialization(); 6421 } 6422 6423 private Builder( 6424 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6425 super(parent); 6426 maybeForceBuilderInitialization(); 6427 } 6428 private void maybeForceBuilderInitialization() { 6429 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6430 getReqInfoFieldBuilder(); 6431 } 6432 } 6433 private static Builder create() { 6434 return new Builder(); 6435 } 6436 6437 public Builder clear() { 6438 super.clear(); 6439 if (reqInfoBuilder_ == null) { 6440 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 6441 } else { 6442 reqInfoBuilder_.clear(); 6443 } 6444 bitField0_ = (bitField0_ & ~0x00000001); 6445 startTxId_ = 0L; 6446 bitField0_ = (bitField0_ & ~0x00000002); 6447 endTxId_ = 0L; 6448 bitField0_ = (bitField0_ & ~0x00000004); 6449 return this; 6450 } 6451 6452 public Builder clone() { 6453 return create().mergeFrom(buildPartial()); 6454 } 6455 6456 public com.google.protobuf.Descriptors.Descriptor 6457 getDescriptorForType() { 6458 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor; 6459 } 6460 6461 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto getDefaultInstanceForType() { 6462 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance(); 6463 } 6464 6465 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto build() { 6466 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = buildPartial(); 6467 if (!result.isInitialized()) { 6468 throw newUninitializedMessageException(result); 6469 } 6470 return result; 6471 } 6472 6473 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto buildPartial() { 6474 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto(this); 6475 int from_bitField0_ = bitField0_; 6476 int to_bitField0_ = 0; 6477 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 6478 to_bitField0_ |= 0x00000001; 6479 } 6480 if (reqInfoBuilder_ == null) { 6481 result.reqInfo_ = reqInfo_; 6482 } else { 6483 result.reqInfo_ = reqInfoBuilder_.build(); 6484 } 6485 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 6486 to_bitField0_ |= 0x00000002; 6487 } 6488 result.startTxId_ = startTxId_; 6489 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 6490 to_bitField0_ |= 0x00000004; 6491 } 6492 result.endTxId_ = endTxId_; 6493 result.bitField0_ = to_bitField0_; 6494 onBuilt(); 6495 return result; 6496 } 6497 6498 public Builder mergeFrom(com.google.protobuf.Message other) { 6499 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) { 6500 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)other); 6501 } else { 6502 super.mergeFrom(other); 6503 return this; 6504 } 6505 } 6506 6507 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other) { 6508 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance()) return this; 6509 if (other.hasReqInfo()) { 6510 mergeReqInfo(other.getReqInfo()); 6511 } 6512 if (other.hasStartTxId()) { 6513 setStartTxId(other.getStartTxId()); 6514 } 6515 if (other.hasEndTxId()) { 6516 setEndTxId(other.getEndTxId()); 6517 } 6518 this.mergeUnknownFields(other.getUnknownFields()); 6519 return this; 6520 } 6521 6522 public final boolean isInitialized() { 6523 if (!hasReqInfo()) { 6524 6525 return false; 6526 } 6527 if (!hasStartTxId()) { 6528 6529 return false; 6530 } 6531 if (!hasEndTxId()) { 6532 6533 return false; 6534 } 6535 if (!getReqInfo().isInitialized()) { 6536 6537 return false; 6538 } 6539 return true; 6540 } 6541 6542 public Builder mergeFrom( 6543 com.google.protobuf.CodedInputStream input, 6544 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6545 throws java.io.IOException { 6546 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parsedMessage = null; 6547 try { 6548 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6549 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6550 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) e.getUnfinishedMessage(); 6551 throw e; 6552 } finally { 6553 if (parsedMessage != null) { 6554 mergeFrom(parsedMessage); 6555 } 6556 } 6557 return this; 6558 } 6559 private int bitField0_; 6560 6561 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 6562 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 6563 private com.google.protobuf.SingleFieldBuilder< 6564 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 6565 /** 6566 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6567 */ 6568 public boolean hasReqInfo() { 6569 return ((bitField0_ & 0x00000001) == 0x00000001); 6570 } 6571 /** 6572 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6573 */ 6574 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 6575 if (reqInfoBuilder_ == null) { 6576 return reqInfo_; 6577 } else { 6578 return reqInfoBuilder_.getMessage(); 6579 } 6580 } 6581 /** 6582 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6583 */ 6584 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 6585 if (reqInfoBuilder_ == null) { 6586 if (value == null) { 6587 throw new NullPointerException(); 6588 } 6589 reqInfo_ = value; 6590 onChanged(); 6591 } else { 6592 reqInfoBuilder_.setMessage(value); 6593 } 6594 bitField0_ |= 0x00000001; 6595 return this; 6596 } 6597 /** 6598 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6599 */ 6600 public Builder setReqInfo( 6601 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 6602 if (reqInfoBuilder_ == null) { 6603 reqInfo_ = builderForValue.build(); 6604 onChanged(); 6605 } else { 6606 reqInfoBuilder_.setMessage(builderForValue.build()); 6607 } 6608 bitField0_ |= 0x00000001; 6609 return this; 6610 } 6611 /** 6612 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6613 */ 6614 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 6615 if (reqInfoBuilder_ == null) { 6616 if (((bitField0_ & 0x00000001) == 0x00000001) && 6617 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 6618 reqInfo_ = 6619 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 6620 } else { 6621 reqInfo_ = value; 6622 } 6623 onChanged(); 6624 } else { 6625 reqInfoBuilder_.mergeFrom(value); 6626 } 6627 bitField0_ |= 0x00000001; 6628 return this; 6629 } 6630 /** 6631 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6632 */ 6633 public Builder clearReqInfo() { 6634 if (reqInfoBuilder_ == null) { 6635 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 6636 onChanged(); 6637 } else { 6638 reqInfoBuilder_.clear(); 6639 } 6640 bitField0_ = (bitField0_ & ~0x00000001); 6641 return this; 6642 } 6643 /** 6644 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6645 */ 6646 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 6647 bitField0_ |= 0x00000001; 6648 onChanged(); 6649 return getReqInfoFieldBuilder().getBuilder(); 6650 } 6651 /** 6652 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6653 */ 6654 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 6655 if (reqInfoBuilder_ != null) { 6656 return reqInfoBuilder_.getMessageOrBuilder(); 6657 } else { 6658 return reqInfo_; 6659 } 6660 } 6661 /** 6662 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 6663 */ 6664 private com.google.protobuf.SingleFieldBuilder< 6665 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 6666 getReqInfoFieldBuilder() { 6667 if (reqInfoBuilder_ == null) { 6668 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 6669 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 6670 reqInfo_, 6671 getParentForChildren(), 6672 isClean()); 6673 reqInfo_ = null; 6674 } 6675 return reqInfoBuilder_; 6676 } 6677 6678 // required uint64 startTxId = 2; 6679 private long startTxId_ ; 6680 /** 6681 * <code>required uint64 startTxId = 2;</code> 6682 */ 6683 public boolean hasStartTxId() { 6684 return ((bitField0_ & 0x00000002) == 0x00000002); 6685 } 6686 /** 6687 * <code>required uint64 startTxId = 2;</code> 6688 */ 6689 public long getStartTxId() { 6690 return startTxId_; 6691 } 6692 /** 6693 * <code>required uint64 startTxId = 2;</code> 6694 */ 6695 public Builder setStartTxId(long value) { 6696 bitField0_ |= 0x00000002; 6697 startTxId_ = value; 6698 onChanged(); 6699 return this; 6700 } 6701 /** 6702 * <code>required uint64 startTxId = 2;</code> 6703 */ 6704 public Builder clearStartTxId() { 6705 bitField0_ = (bitField0_ & ~0x00000002); 6706 startTxId_ = 0L; 6707 onChanged(); 6708 return this; 6709 } 6710 6711 // required uint64 endTxId = 3; 6712 private long endTxId_ ; 6713 /** 6714 * <code>required uint64 endTxId = 3;</code> 6715 */ 6716 public boolean hasEndTxId() { 6717 return ((bitField0_ & 0x00000004) == 0x00000004); 6718 } 6719 /** 6720 * <code>required uint64 endTxId = 3;</code> 6721 */ 6722 public long getEndTxId() { 6723 return endTxId_; 6724 } 6725 /** 6726 * <code>required uint64 endTxId = 3;</code> 6727 */ 6728 public Builder setEndTxId(long value) { 6729 bitField0_ |= 0x00000004; 6730 endTxId_ = value; 6731 onChanged(); 6732 return this; 6733 } 6734 /** 6735 * <code>required uint64 endTxId = 3;</code> 6736 */ 6737 public Builder clearEndTxId() { 6738 bitField0_ = (bitField0_ & ~0x00000004); 6739 endTxId_ = 0L; 6740 onChanged(); 6741 return this; 6742 } 6743 6744 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto) 6745 } 6746 6747 static { 6748 defaultInstance = new FinalizeLogSegmentRequestProto(true); 6749 defaultInstance.initFields(); 6750 } 6751 6752 // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto) 6753 } 6754 6755 public interface FinalizeLogSegmentResponseProtoOrBuilder 6756 extends com.google.protobuf.MessageOrBuilder { 6757 } 6758 /** 6759 * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentResponseProto} 6760 */ 6761 public static final class FinalizeLogSegmentResponseProto extends 6762 com.google.protobuf.GeneratedMessage 6763 implements FinalizeLogSegmentResponseProtoOrBuilder { 6764 // Use FinalizeLogSegmentResponseProto.newBuilder() to construct. 6765 private FinalizeLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 6766 super(builder); 6767 this.unknownFields = builder.getUnknownFields(); 6768 } 6769 private FinalizeLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 6770 6771 private static final FinalizeLogSegmentResponseProto defaultInstance; 6772 public static FinalizeLogSegmentResponseProto getDefaultInstance() { 6773 return defaultInstance; 6774 } 6775 6776 public FinalizeLogSegmentResponseProto getDefaultInstanceForType() { 6777 return defaultInstance; 6778 } 6779 6780 private final com.google.protobuf.UnknownFieldSet unknownFields; 6781 @java.lang.Override 6782 public final com.google.protobuf.UnknownFieldSet 6783 getUnknownFields() { 6784 return this.unknownFields; 6785 } 6786 private FinalizeLogSegmentResponseProto( 6787 com.google.protobuf.CodedInputStream input, 6788 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6789 throws com.google.protobuf.InvalidProtocolBufferException { 6790 initFields(); 6791 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 6792 com.google.protobuf.UnknownFieldSet.newBuilder(); 6793 try { 6794 boolean done = false; 6795 while (!done) { 6796 int tag = input.readTag(); 6797 switch (tag) { 6798 case 0: 6799 done = true; 6800 break; 6801 default: { 6802 if (!parseUnknownField(input, unknownFields, 6803 extensionRegistry, tag)) { 6804 done = true; 6805 } 6806 break; 6807 } 6808 } 6809 } 6810 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6811 throw e.setUnfinishedMessage(this); 6812 } catch (java.io.IOException e) { 6813 throw new com.google.protobuf.InvalidProtocolBufferException( 6814 e.getMessage()).setUnfinishedMessage(this); 6815 } finally { 6816 this.unknownFields = unknownFields.build(); 6817 makeExtensionsImmutable(); 6818 } 6819 } 6820 public static final com.google.protobuf.Descriptors.Descriptor 6821 getDescriptor() { 6822 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor; 6823 } 6824 6825 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6826 internalGetFieldAccessorTable() { 6827 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable 6828 .ensureFieldAccessorsInitialized( 6829 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class); 6830 } 6831 6832 public static com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> PARSER = 6833 new com.google.protobuf.AbstractParser<FinalizeLogSegmentResponseProto>() { 6834 public FinalizeLogSegmentResponseProto parsePartialFrom( 6835 com.google.protobuf.CodedInputStream input, 6836 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6837 throws com.google.protobuf.InvalidProtocolBufferException { 6838 return new FinalizeLogSegmentResponseProto(input, extensionRegistry); 6839 } 6840 }; 6841 6842 @java.lang.Override 6843 public com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> getParserForType() { 6844 return PARSER; 6845 } 6846 6847 private void initFields() { 6848 } 6849 private byte memoizedIsInitialized = -1; 6850 public final boolean isInitialized() { 6851 byte isInitialized = memoizedIsInitialized; 6852 if (isInitialized != -1) return isInitialized == 1; 6853 6854 memoizedIsInitialized = 1; 6855 return true; 6856 } 6857 6858 public void writeTo(com.google.protobuf.CodedOutputStream output) 6859 throws java.io.IOException { 6860 getSerializedSize(); 6861 getUnknownFields().writeTo(output); 6862 } 6863 6864 private int memoizedSerializedSize = -1; 6865 public int getSerializedSize() { 6866 int size = memoizedSerializedSize; 6867 if (size != -1) return size; 6868 6869 size = 0; 6870 size += getUnknownFields().getSerializedSize(); 6871 memoizedSerializedSize = size; 6872 return size; 6873 } 6874 6875 private static final long serialVersionUID = 0L; 6876 @java.lang.Override 6877 protected java.lang.Object writeReplace() 6878 throws java.io.ObjectStreamException { 6879 return super.writeReplace(); 6880 } 6881 6882 @java.lang.Override 6883 public boolean equals(final java.lang.Object obj) { 6884 if (obj == this) { 6885 return true; 6886 } 6887 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)) { 6888 return super.equals(obj); 6889 } 6890 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) obj; 6891 6892 boolean result = true; 6893 result = result && 6894 getUnknownFields().equals(other.getUnknownFields()); 6895 return result; 6896 } 6897 6898 private int memoizedHashCode = 0; 6899 @java.lang.Override 6900 public int hashCode() { 6901 if (memoizedHashCode != 0) { 6902 return memoizedHashCode; 6903 } 6904 int hash = 41; 6905 hash = (19 * hash) + getDescriptorForType().hashCode(); 6906 hash = (29 * hash) + getUnknownFields().hashCode(); 6907 memoizedHashCode = hash; 6908 return hash; 6909 } 6910 6911 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6912 com.google.protobuf.ByteString data) 6913 throws com.google.protobuf.InvalidProtocolBufferException { 6914 return PARSER.parseFrom(data); 6915 } 6916 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6917 com.google.protobuf.ByteString data, 6918 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6919 throws com.google.protobuf.InvalidProtocolBufferException { 6920 return PARSER.parseFrom(data, extensionRegistry); 6921 } 6922 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(byte[] data) 6923 throws com.google.protobuf.InvalidProtocolBufferException { 6924 return PARSER.parseFrom(data); 6925 } 6926 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6927 byte[] data, 6928 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6929 throws com.google.protobuf.InvalidProtocolBufferException { 6930 return PARSER.parseFrom(data, extensionRegistry); 6931 } 6932 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(java.io.InputStream input) 6933 throws java.io.IOException { 6934 return PARSER.parseFrom(input); 6935 } 6936 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6937 java.io.InputStream input, 6938 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6939 throws java.io.IOException { 6940 return PARSER.parseFrom(input, extensionRegistry); 6941 } 6942 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input) 6943 throws java.io.IOException { 6944 return PARSER.parseDelimitedFrom(input); 6945 } 6946 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom( 6947 java.io.InputStream input, 6948 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6949 throws java.io.IOException { 6950 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6951 } 6952 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6953 com.google.protobuf.CodedInputStream input) 6954 throws java.io.IOException { 6955 return PARSER.parseFrom(input); 6956 } 6957 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom( 6958 com.google.protobuf.CodedInputStream input, 6959 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6960 throws java.io.IOException { 6961 return PARSER.parseFrom(input, extensionRegistry); 6962 } 6963 6964 public static Builder newBuilder() { return Builder.create(); } 6965 public Builder newBuilderForType() { return newBuilder(); } 6966 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto prototype) { 6967 return newBuilder().mergeFrom(prototype); 6968 } 6969 public Builder toBuilder() { return newBuilder(this); } 6970 6971 @java.lang.Override 6972 protected Builder newBuilderForType( 6973 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6974 Builder builder = new Builder(parent); 6975 return builder; 6976 } 6977 /** 6978 * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentResponseProto} 6979 */ 6980 public static final class Builder extends 6981 com.google.protobuf.GeneratedMessage.Builder<Builder> 6982 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProtoOrBuilder { 6983 public static final com.google.protobuf.Descriptors.Descriptor 6984 getDescriptor() { 6985 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor; 6986 } 6987 6988 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6989 internalGetFieldAccessorTable() { 6990 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable 6991 .ensureFieldAccessorsInitialized( 6992 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class); 6993 } 6994 6995 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.newBuilder() 6996 private Builder() { 6997 maybeForceBuilderInitialization(); 6998 } 6999 7000 private Builder( 7001 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7002 super(parent); 7003 maybeForceBuilderInitialization(); 7004 } 7005 private void maybeForceBuilderInitialization() { 7006 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7007 } 7008 } 7009 private static Builder create() { 7010 return new Builder(); 7011 } 7012 7013 public Builder clear() { 7014 super.clear(); 7015 return this; 7016 } 7017 7018 public Builder clone() { 7019 return create().mergeFrom(buildPartial()); 7020 } 7021 7022 public com.google.protobuf.Descriptors.Descriptor 7023 getDescriptorForType() { 7024 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor; 7025 } 7026 7027 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto getDefaultInstanceForType() { 7028 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(); 7029 } 7030 7031 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto build() { 7032 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = buildPartial(); 7033 if (!result.isInitialized()) { 7034 throw newUninitializedMessageException(result); 7035 } 7036 return result; 7037 } 7038 7039 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto buildPartial() { 7040 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto(this); 7041 onBuilt(); 7042 return result; 7043 } 7044 7045 public Builder mergeFrom(com.google.protobuf.Message other) { 7046 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) { 7047 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)other); 7048 } else { 7049 super.mergeFrom(other); 7050 return this; 7051 } 7052 } 7053 7054 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other) { 7055 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()) return this; 7056 this.mergeUnknownFields(other.getUnknownFields()); 7057 return this; 7058 } 7059 7060 public final boolean isInitialized() { 7061 return true; 7062 } 7063 7064 public Builder mergeFrom( 7065 com.google.protobuf.CodedInputStream input, 7066 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7067 throws java.io.IOException { 7068 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parsedMessage = null; 7069 try { 7070 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7071 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7072 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) e.getUnfinishedMessage(); 7073 throw e; 7074 } finally { 7075 if (parsedMessage != null) { 7076 mergeFrom(parsedMessage); 7077 } 7078 } 7079 return this; 7080 } 7081 7082 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto) 7083 } 7084 7085 static { 7086 defaultInstance = new FinalizeLogSegmentResponseProto(true); 7087 defaultInstance.initFields(); 7088 } 7089 7090 // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto) 7091 } 7092 7093 public interface PurgeLogsRequestProtoOrBuilder 7094 extends com.google.protobuf.MessageOrBuilder { 7095 7096 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 7097 /** 7098 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7099 */ 7100 boolean hasReqInfo(); 7101 /** 7102 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7103 */ 7104 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 7105 /** 7106 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7107 */ 7108 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 7109 7110 // required uint64 minTxIdToKeep = 2; 7111 /** 7112 * <code>required uint64 minTxIdToKeep = 2;</code> 7113 */ 7114 boolean hasMinTxIdToKeep(); 7115 /** 7116 * <code>required uint64 minTxIdToKeep = 2;</code> 7117 */ 7118 long getMinTxIdToKeep(); 7119 } 7120 /** 7121 * Protobuf type {@code hadoop.hdfs.PurgeLogsRequestProto} 7122 * 7123 * <pre> 7124 ** 7125 * purgeLogs() 7126 * </pre> 7127 */ 7128 public static final class PurgeLogsRequestProto extends 7129 com.google.protobuf.GeneratedMessage 7130 implements PurgeLogsRequestProtoOrBuilder { 7131 // Use PurgeLogsRequestProto.newBuilder() to construct. 7132 private PurgeLogsRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7133 super(builder); 7134 this.unknownFields = builder.getUnknownFields(); 7135 } 7136 private PurgeLogsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7137 7138 private static final PurgeLogsRequestProto defaultInstance; 7139 public static PurgeLogsRequestProto getDefaultInstance() { 7140 return defaultInstance; 7141 } 7142 7143 public PurgeLogsRequestProto getDefaultInstanceForType() { 7144 return defaultInstance; 7145 } 7146 7147 private final com.google.protobuf.UnknownFieldSet unknownFields; 7148 @java.lang.Override 7149 public final com.google.protobuf.UnknownFieldSet 7150 getUnknownFields() { 7151 return this.unknownFields; 7152 } 7153 private PurgeLogsRequestProto( 7154 com.google.protobuf.CodedInputStream input, 7155 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7156 throws com.google.protobuf.InvalidProtocolBufferException { 7157 initFields(); 7158 int mutable_bitField0_ = 0; 7159 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7160 com.google.protobuf.UnknownFieldSet.newBuilder(); 7161 try { 7162 boolean done = false; 7163 while (!done) { 7164 int tag = input.readTag(); 7165 switch (tag) { 7166 case 0: 7167 done = true; 7168 break; 7169 default: { 7170 if (!parseUnknownField(input, unknownFields, 7171 extensionRegistry, tag)) { 7172 done = true; 7173 } 7174 break; 7175 } 7176 case 10: { 7177 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 7178 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7179 subBuilder = reqInfo_.toBuilder(); 7180 } 7181 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 7182 if (subBuilder != null) { 7183 subBuilder.mergeFrom(reqInfo_); 7184 reqInfo_ = subBuilder.buildPartial(); 7185 } 7186 bitField0_ |= 0x00000001; 7187 break; 7188 } 7189 case 16: { 7190 bitField0_ |= 0x00000002; 7191 minTxIdToKeep_ = input.readUInt64(); 7192 break; 7193 } 7194 } 7195 } 7196 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7197 throw e.setUnfinishedMessage(this); 7198 } catch (java.io.IOException e) { 7199 throw new com.google.protobuf.InvalidProtocolBufferException( 7200 e.getMessage()).setUnfinishedMessage(this); 7201 } finally { 7202 this.unknownFields = unknownFields.build(); 7203 makeExtensionsImmutable(); 7204 } 7205 } 7206 public static final com.google.protobuf.Descriptors.Descriptor 7207 getDescriptor() { 7208 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor; 7209 } 7210 7211 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7212 internalGetFieldAccessorTable() { 7213 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable 7214 .ensureFieldAccessorsInitialized( 7215 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class); 7216 } 7217 7218 public static com.google.protobuf.Parser<PurgeLogsRequestProto> PARSER = 7219 new com.google.protobuf.AbstractParser<PurgeLogsRequestProto>() { 7220 public PurgeLogsRequestProto parsePartialFrom( 7221 com.google.protobuf.CodedInputStream input, 7222 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7223 throws com.google.protobuf.InvalidProtocolBufferException { 7224 return new PurgeLogsRequestProto(input, extensionRegistry); 7225 } 7226 }; 7227 7228 @java.lang.Override 7229 public com.google.protobuf.Parser<PurgeLogsRequestProto> getParserForType() { 7230 return PARSER; 7231 } 7232 7233 private int bitField0_; 7234 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 7235 public static final int REQINFO_FIELD_NUMBER = 1; 7236 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 7237 /** 7238 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7239 */ 7240 public boolean hasReqInfo() { 7241 return ((bitField0_ & 0x00000001) == 0x00000001); 7242 } 7243 /** 7244 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7245 */ 7246 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 7247 return reqInfo_; 7248 } 7249 /** 7250 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7251 */ 7252 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 7253 return reqInfo_; 7254 } 7255 7256 // required uint64 minTxIdToKeep = 2; 7257 public static final int MINTXIDTOKEEP_FIELD_NUMBER = 2; 7258 private long minTxIdToKeep_; 7259 /** 7260 * <code>required uint64 minTxIdToKeep = 2;</code> 7261 */ 7262 public boolean hasMinTxIdToKeep() { 7263 return ((bitField0_ & 0x00000002) == 0x00000002); 7264 } 7265 /** 7266 * <code>required uint64 minTxIdToKeep = 2;</code> 7267 */ 7268 public long getMinTxIdToKeep() { 7269 return minTxIdToKeep_; 7270 } 7271 7272 private void initFields() { 7273 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 7274 minTxIdToKeep_ = 0L; 7275 } 7276 private byte memoizedIsInitialized = -1; 7277 public final boolean isInitialized() { 7278 byte isInitialized = memoizedIsInitialized; 7279 if (isInitialized != -1) return isInitialized == 1; 7280 7281 if (!hasReqInfo()) { 7282 memoizedIsInitialized = 0; 7283 return false; 7284 } 7285 if (!hasMinTxIdToKeep()) { 7286 memoizedIsInitialized = 0; 7287 return false; 7288 } 7289 if (!getReqInfo().isInitialized()) { 7290 memoizedIsInitialized = 0; 7291 return false; 7292 } 7293 memoizedIsInitialized = 1; 7294 return true; 7295 } 7296 7297 public void writeTo(com.google.protobuf.CodedOutputStream output) 7298 throws java.io.IOException { 7299 getSerializedSize(); 7300 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7301 output.writeMessage(1, reqInfo_); 7302 } 7303 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7304 output.writeUInt64(2, minTxIdToKeep_); 7305 } 7306 getUnknownFields().writeTo(output); 7307 } 7308 7309 private int memoizedSerializedSize = -1; 7310 public int getSerializedSize() { 7311 int size = memoizedSerializedSize; 7312 if (size != -1) return size; 7313 7314 size = 0; 7315 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7316 size += com.google.protobuf.CodedOutputStream 7317 .computeMessageSize(1, reqInfo_); 7318 } 7319 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7320 size += com.google.protobuf.CodedOutputStream 7321 .computeUInt64Size(2, minTxIdToKeep_); 7322 } 7323 size += getUnknownFields().getSerializedSize(); 7324 memoizedSerializedSize = size; 7325 return size; 7326 } 7327 7328 private static final long serialVersionUID = 0L; 7329 @java.lang.Override 7330 protected java.lang.Object writeReplace() 7331 throws java.io.ObjectStreamException { 7332 return super.writeReplace(); 7333 } 7334 7335 @java.lang.Override 7336 public boolean equals(final java.lang.Object obj) { 7337 if (obj == this) { 7338 return true; 7339 } 7340 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)) { 7341 return super.equals(obj); 7342 } 7343 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) obj; 7344 7345 boolean result = true; 7346 result = result && (hasReqInfo() == other.hasReqInfo()); 7347 if (hasReqInfo()) { 7348 result = result && getReqInfo() 7349 .equals(other.getReqInfo()); 7350 } 7351 result = result && (hasMinTxIdToKeep() == other.hasMinTxIdToKeep()); 7352 if (hasMinTxIdToKeep()) { 7353 result = result && (getMinTxIdToKeep() 7354 == other.getMinTxIdToKeep()); 7355 } 7356 result = result && 7357 getUnknownFields().equals(other.getUnknownFields()); 7358 return result; 7359 } 7360 7361 private int memoizedHashCode = 0; 7362 @java.lang.Override 7363 public int hashCode() { 7364 if (memoizedHashCode != 0) { 7365 return memoizedHashCode; 7366 } 7367 int hash = 41; 7368 hash = (19 * hash) + getDescriptorForType().hashCode(); 7369 if (hasReqInfo()) { 7370 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 7371 hash = (53 * hash) + getReqInfo().hashCode(); 7372 } 7373 if (hasMinTxIdToKeep()) { 7374 hash = (37 * hash) + MINTXIDTOKEEP_FIELD_NUMBER; 7375 hash = (53 * hash) + hashLong(getMinTxIdToKeep()); 7376 } 7377 hash = (29 * hash) + getUnknownFields().hashCode(); 7378 memoizedHashCode = hash; 7379 return hash; 7380 } 7381 7382 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7383 com.google.protobuf.ByteString data) 7384 throws com.google.protobuf.InvalidProtocolBufferException { 7385 return PARSER.parseFrom(data); 7386 } 7387 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7388 com.google.protobuf.ByteString data, 7389 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7390 throws com.google.protobuf.InvalidProtocolBufferException { 7391 return PARSER.parseFrom(data, extensionRegistry); 7392 } 7393 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(byte[] data) 7394 throws com.google.protobuf.InvalidProtocolBufferException { 7395 return PARSER.parseFrom(data); 7396 } 7397 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7398 byte[] data, 7399 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7400 throws com.google.protobuf.InvalidProtocolBufferException { 7401 return PARSER.parseFrom(data, extensionRegistry); 7402 } 7403 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(java.io.InputStream input) 7404 throws java.io.IOException { 7405 return PARSER.parseFrom(input); 7406 } 7407 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7408 java.io.InputStream input, 7409 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7410 throws java.io.IOException { 7411 return PARSER.parseFrom(input, extensionRegistry); 7412 } 7413 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(java.io.InputStream input) 7414 throws java.io.IOException { 7415 return PARSER.parseDelimitedFrom(input); 7416 } 7417 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom( 7418 java.io.InputStream input, 7419 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7420 throws java.io.IOException { 7421 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7422 } 7423 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7424 com.google.protobuf.CodedInputStream input) 7425 throws java.io.IOException { 7426 return PARSER.parseFrom(input); 7427 } 7428 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom( 7429 com.google.protobuf.CodedInputStream input, 7430 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7431 throws java.io.IOException { 7432 return PARSER.parseFrom(input, extensionRegistry); 7433 } 7434 7435 public static Builder newBuilder() { return Builder.create(); } 7436 public Builder newBuilderForType() { return newBuilder(); } 7437 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto prototype) { 7438 return newBuilder().mergeFrom(prototype); 7439 } 7440 public Builder toBuilder() { return newBuilder(this); } 7441 7442 @java.lang.Override 7443 protected Builder newBuilderForType( 7444 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7445 Builder builder = new Builder(parent); 7446 return builder; 7447 } 7448 /** 7449 * Protobuf type {@code hadoop.hdfs.PurgeLogsRequestProto} 7450 * 7451 * <pre> 7452 ** 7453 * purgeLogs() 7454 * </pre> 7455 */ 7456 public static final class Builder extends 7457 com.google.protobuf.GeneratedMessage.Builder<Builder> 7458 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProtoOrBuilder { 7459 public static final com.google.protobuf.Descriptors.Descriptor 7460 getDescriptor() { 7461 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor; 7462 } 7463 7464 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7465 internalGetFieldAccessorTable() { 7466 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable 7467 .ensureFieldAccessorsInitialized( 7468 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class); 7469 } 7470 7471 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.newBuilder() 7472 private Builder() { 7473 maybeForceBuilderInitialization(); 7474 } 7475 7476 private Builder( 7477 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7478 super(parent); 7479 maybeForceBuilderInitialization(); 7480 } 7481 private void maybeForceBuilderInitialization() { 7482 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7483 getReqInfoFieldBuilder(); 7484 } 7485 } 7486 private static Builder create() { 7487 return new Builder(); 7488 } 7489 7490 public Builder clear() { 7491 super.clear(); 7492 if (reqInfoBuilder_ == null) { 7493 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 7494 } else { 7495 reqInfoBuilder_.clear(); 7496 } 7497 bitField0_ = (bitField0_ & ~0x00000001); 7498 minTxIdToKeep_ = 0L; 7499 bitField0_ = (bitField0_ & ~0x00000002); 7500 return this; 7501 } 7502 7503 public Builder clone() { 7504 return create().mergeFrom(buildPartial()); 7505 } 7506 7507 public com.google.protobuf.Descriptors.Descriptor 7508 getDescriptorForType() { 7509 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor; 7510 } 7511 7512 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto getDefaultInstanceForType() { 7513 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance(); 7514 } 7515 7516 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto build() { 7517 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = buildPartial(); 7518 if (!result.isInitialized()) { 7519 throw newUninitializedMessageException(result); 7520 } 7521 return result; 7522 } 7523 7524 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto buildPartial() { 7525 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto(this); 7526 int from_bitField0_ = bitField0_; 7527 int to_bitField0_ = 0; 7528 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 7529 to_bitField0_ |= 0x00000001; 7530 } 7531 if (reqInfoBuilder_ == null) { 7532 result.reqInfo_ = reqInfo_; 7533 } else { 7534 result.reqInfo_ = reqInfoBuilder_.build(); 7535 } 7536 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 7537 to_bitField0_ |= 0x00000002; 7538 } 7539 result.minTxIdToKeep_ = minTxIdToKeep_; 7540 result.bitField0_ = to_bitField0_; 7541 onBuilt(); 7542 return result; 7543 } 7544 7545 public Builder mergeFrom(com.google.protobuf.Message other) { 7546 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) { 7547 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)other); 7548 } else { 7549 super.mergeFrom(other); 7550 return this; 7551 } 7552 } 7553 7554 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other) { 7555 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance()) return this; 7556 if (other.hasReqInfo()) { 7557 mergeReqInfo(other.getReqInfo()); 7558 } 7559 if (other.hasMinTxIdToKeep()) { 7560 setMinTxIdToKeep(other.getMinTxIdToKeep()); 7561 } 7562 this.mergeUnknownFields(other.getUnknownFields()); 7563 return this; 7564 } 7565 7566 public final boolean isInitialized() { 7567 if (!hasReqInfo()) { 7568 7569 return false; 7570 } 7571 if (!hasMinTxIdToKeep()) { 7572 7573 return false; 7574 } 7575 if (!getReqInfo().isInitialized()) { 7576 7577 return false; 7578 } 7579 return true; 7580 } 7581 7582 public Builder mergeFrom( 7583 com.google.protobuf.CodedInputStream input, 7584 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7585 throws java.io.IOException { 7586 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parsedMessage = null; 7587 try { 7588 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7589 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7590 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) e.getUnfinishedMessage(); 7591 throw e; 7592 } finally { 7593 if (parsedMessage != null) { 7594 mergeFrom(parsedMessage); 7595 } 7596 } 7597 return this; 7598 } 7599 private int bitField0_; 7600 7601 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 7602 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 7603 private com.google.protobuf.SingleFieldBuilder< 7604 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 7605 /** 7606 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7607 */ 7608 public boolean hasReqInfo() { 7609 return ((bitField0_ & 0x00000001) == 0x00000001); 7610 } 7611 /** 7612 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7613 */ 7614 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 7615 if (reqInfoBuilder_ == null) { 7616 return reqInfo_; 7617 } else { 7618 return reqInfoBuilder_.getMessage(); 7619 } 7620 } 7621 /** 7622 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7623 */ 7624 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 7625 if (reqInfoBuilder_ == null) { 7626 if (value == null) { 7627 throw new NullPointerException(); 7628 } 7629 reqInfo_ = value; 7630 onChanged(); 7631 } else { 7632 reqInfoBuilder_.setMessage(value); 7633 } 7634 bitField0_ |= 0x00000001; 7635 return this; 7636 } 7637 /** 7638 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7639 */ 7640 public Builder setReqInfo( 7641 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 7642 if (reqInfoBuilder_ == null) { 7643 reqInfo_ = builderForValue.build(); 7644 onChanged(); 7645 } else { 7646 reqInfoBuilder_.setMessage(builderForValue.build()); 7647 } 7648 bitField0_ |= 0x00000001; 7649 return this; 7650 } 7651 /** 7652 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7653 */ 7654 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 7655 if (reqInfoBuilder_ == null) { 7656 if (((bitField0_ & 0x00000001) == 0x00000001) && 7657 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 7658 reqInfo_ = 7659 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 7660 } else { 7661 reqInfo_ = value; 7662 } 7663 onChanged(); 7664 } else { 7665 reqInfoBuilder_.mergeFrom(value); 7666 } 7667 bitField0_ |= 0x00000001; 7668 return this; 7669 } 7670 /** 7671 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7672 */ 7673 public Builder clearReqInfo() { 7674 if (reqInfoBuilder_ == null) { 7675 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 7676 onChanged(); 7677 } else { 7678 reqInfoBuilder_.clear(); 7679 } 7680 bitField0_ = (bitField0_ & ~0x00000001); 7681 return this; 7682 } 7683 /** 7684 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7685 */ 7686 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 7687 bitField0_ |= 0x00000001; 7688 onChanged(); 7689 return getReqInfoFieldBuilder().getBuilder(); 7690 } 7691 /** 7692 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7693 */ 7694 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 7695 if (reqInfoBuilder_ != null) { 7696 return reqInfoBuilder_.getMessageOrBuilder(); 7697 } else { 7698 return reqInfo_; 7699 } 7700 } 7701 /** 7702 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 7703 */ 7704 private com.google.protobuf.SingleFieldBuilder< 7705 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 7706 getReqInfoFieldBuilder() { 7707 if (reqInfoBuilder_ == null) { 7708 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 7709 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 7710 reqInfo_, 7711 getParentForChildren(), 7712 isClean()); 7713 reqInfo_ = null; 7714 } 7715 return reqInfoBuilder_; 7716 } 7717 7718 // required uint64 minTxIdToKeep = 2; 7719 private long minTxIdToKeep_ ; 7720 /** 7721 * <code>required uint64 minTxIdToKeep = 2;</code> 7722 */ 7723 public boolean hasMinTxIdToKeep() { 7724 return ((bitField0_ & 0x00000002) == 0x00000002); 7725 } 7726 /** 7727 * <code>required uint64 minTxIdToKeep = 2;</code> 7728 */ 7729 public long getMinTxIdToKeep() { 7730 return minTxIdToKeep_; 7731 } 7732 /** 7733 * <code>required uint64 minTxIdToKeep = 2;</code> 7734 */ 7735 public Builder setMinTxIdToKeep(long value) { 7736 bitField0_ |= 0x00000002; 7737 minTxIdToKeep_ = value; 7738 onChanged(); 7739 return this; 7740 } 7741 /** 7742 * <code>required uint64 minTxIdToKeep = 2;</code> 7743 */ 7744 public Builder clearMinTxIdToKeep() { 7745 bitField0_ = (bitField0_ & ~0x00000002); 7746 minTxIdToKeep_ = 0L; 7747 onChanged(); 7748 return this; 7749 } 7750 7751 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsRequestProto) 7752 } 7753 7754 static { 7755 defaultInstance = new PurgeLogsRequestProto(true); 7756 defaultInstance.initFields(); 7757 } 7758 7759 // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsRequestProto) 7760 } 7761 7762 public interface PurgeLogsResponseProtoOrBuilder 7763 extends com.google.protobuf.MessageOrBuilder { 7764 } 7765 /** 7766 * Protobuf type {@code hadoop.hdfs.PurgeLogsResponseProto} 7767 */ 7768 public static final class PurgeLogsResponseProto extends 7769 com.google.protobuf.GeneratedMessage 7770 implements PurgeLogsResponseProtoOrBuilder { 7771 // Use PurgeLogsResponseProto.newBuilder() to construct. 7772 private PurgeLogsResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7773 super(builder); 7774 this.unknownFields = builder.getUnknownFields(); 7775 } 7776 private PurgeLogsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7777 7778 private static final PurgeLogsResponseProto defaultInstance; 7779 public static PurgeLogsResponseProto getDefaultInstance() { 7780 return defaultInstance; 7781 } 7782 7783 public PurgeLogsResponseProto getDefaultInstanceForType() { 7784 return defaultInstance; 7785 } 7786 7787 private final com.google.protobuf.UnknownFieldSet unknownFields; 7788 @java.lang.Override 7789 public final com.google.protobuf.UnknownFieldSet 7790 getUnknownFields() { 7791 return this.unknownFields; 7792 } 7793 private PurgeLogsResponseProto( 7794 com.google.protobuf.CodedInputStream input, 7795 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7796 throws com.google.protobuf.InvalidProtocolBufferException { 7797 initFields(); 7798 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7799 com.google.protobuf.UnknownFieldSet.newBuilder(); 7800 try { 7801 boolean done = false; 7802 while (!done) { 7803 int tag = input.readTag(); 7804 switch (tag) { 7805 case 0: 7806 done = true; 7807 break; 7808 default: { 7809 if (!parseUnknownField(input, unknownFields, 7810 extensionRegistry, tag)) { 7811 done = true; 7812 } 7813 break; 7814 } 7815 } 7816 } 7817 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7818 throw e.setUnfinishedMessage(this); 7819 } catch (java.io.IOException e) { 7820 throw new com.google.protobuf.InvalidProtocolBufferException( 7821 e.getMessage()).setUnfinishedMessage(this); 7822 } finally { 7823 this.unknownFields = unknownFields.build(); 7824 makeExtensionsImmutable(); 7825 } 7826 } 7827 public static final com.google.protobuf.Descriptors.Descriptor 7828 getDescriptor() { 7829 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor; 7830 } 7831 7832 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7833 internalGetFieldAccessorTable() { 7834 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable 7835 .ensureFieldAccessorsInitialized( 7836 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class); 7837 } 7838 7839 public static com.google.protobuf.Parser<PurgeLogsResponseProto> PARSER = 7840 new com.google.protobuf.AbstractParser<PurgeLogsResponseProto>() { 7841 public PurgeLogsResponseProto parsePartialFrom( 7842 com.google.protobuf.CodedInputStream input, 7843 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7844 throws com.google.protobuf.InvalidProtocolBufferException { 7845 return new PurgeLogsResponseProto(input, extensionRegistry); 7846 } 7847 }; 7848 7849 @java.lang.Override 7850 public com.google.protobuf.Parser<PurgeLogsResponseProto> getParserForType() { 7851 return PARSER; 7852 } 7853 7854 private void initFields() { 7855 } 7856 private byte memoizedIsInitialized = -1; 7857 public final boolean isInitialized() { 7858 byte isInitialized = memoizedIsInitialized; 7859 if (isInitialized != -1) return isInitialized == 1; 7860 7861 memoizedIsInitialized = 1; 7862 return true; 7863 } 7864 7865 public void writeTo(com.google.protobuf.CodedOutputStream output) 7866 throws java.io.IOException { 7867 getSerializedSize(); 7868 getUnknownFields().writeTo(output); 7869 } 7870 7871 private int memoizedSerializedSize = -1; 7872 public int getSerializedSize() { 7873 int size = memoizedSerializedSize; 7874 if (size != -1) return size; 7875 7876 size = 0; 7877 size += getUnknownFields().getSerializedSize(); 7878 memoizedSerializedSize = size; 7879 return size; 7880 } 7881 7882 private static final long serialVersionUID = 0L; 7883 @java.lang.Override 7884 protected java.lang.Object writeReplace() 7885 throws java.io.ObjectStreamException { 7886 return super.writeReplace(); 7887 } 7888 7889 @java.lang.Override 7890 public boolean equals(final java.lang.Object obj) { 7891 if (obj == this) { 7892 return true; 7893 } 7894 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)) { 7895 return super.equals(obj); 7896 } 7897 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) obj; 7898 7899 boolean result = true; 7900 result = result && 7901 getUnknownFields().equals(other.getUnknownFields()); 7902 return result; 7903 } 7904 7905 private int memoizedHashCode = 0; 7906 @java.lang.Override 7907 public int hashCode() { 7908 if (memoizedHashCode != 0) { 7909 return memoizedHashCode; 7910 } 7911 int hash = 41; 7912 hash = (19 * hash) + getDescriptorForType().hashCode(); 7913 hash = (29 * hash) + getUnknownFields().hashCode(); 7914 memoizedHashCode = hash; 7915 return hash; 7916 } 7917 7918 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7919 com.google.protobuf.ByteString data) 7920 throws com.google.protobuf.InvalidProtocolBufferException { 7921 return PARSER.parseFrom(data); 7922 } 7923 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7924 com.google.protobuf.ByteString data, 7925 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7926 throws com.google.protobuf.InvalidProtocolBufferException { 7927 return PARSER.parseFrom(data, extensionRegistry); 7928 } 7929 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(byte[] data) 7930 throws com.google.protobuf.InvalidProtocolBufferException { 7931 return PARSER.parseFrom(data); 7932 } 7933 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7934 byte[] data, 7935 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7936 throws com.google.protobuf.InvalidProtocolBufferException { 7937 return PARSER.parseFrom(data, extensionRegistry); 7938 } 7939 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(java.io.InputStream input) 7940 throws java.io.IOException { 7941 return PARSER.parseFrom(input); 7942 } 7943 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7944 java.io.InputStream input, 7945 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7946 throws java.io.IOException { 7947 return PARSER.parseFrom(input, extensionRegistry); 7948 } 7949 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(java.io.InputStream input) 7950 throws java.io.IOException { 7951 return PARSER.parseDelimitedFrom(input); 7952 } 7953 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom( 7954 java.io.InputStream input, 7955 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7956 throws java.io.IOException { 7957 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7958 } 7959 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7960 com.google.protobuf.CodedInputStream input) 7961 throws java.io.IOException { 7962 return PARSER.parseFrom(input); 7963 } 7964 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom( 7965 com.google.protobuf.CodedInputStream input, 7966 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7967 throws java.io.IOException { 7968 return PARSER.parseFrom(input, extensionRegistry); 7969 } 7970 7971 public static Builder newBuilder() { return Builder.create(); } 7972 public Builder newBuilderForType() { return newBuilder(); } 7973 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto prototype) { 7974 return newBuilder().mergeFrom(prototype); 7975 } 7976 public Builder toBuilder() { return newBuilder(this); } 7977 7978 @java.lang.Override 7979 protected Builder newBuilderForType( 7980 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7981 Builder builder = new Builder(parent); 7982 return builder; 7983 } 7984 /** 7985 * Protobuf type {@code hadoop.hdfs.PurgeLogsResponseProto} 7986 */ 7987 public static final class Builder extends 7988 com.google.protobuf.GeneratedMessage.Builder<Builder> 7989 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProtoOrBuilder { 7990 public static final com.google.protobuf.Descriptors.Descriptor 7991 getDescriptor() { 7992 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor; 7993 } 7994 7995 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7996 internalGetFieldAccessorTable() { 7997 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable 7998 .ensureFieldAccessorsInitialized( 7999 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class); 8000 } 8001 8002 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.newBuilder() 8003 private Builder() { 8004 maybeForceBuilderInitialization(); 8005 } 8006 8007 private Builder( 8008 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8009 super(parent); 8010 maybeForceBuilderInitialization(); 8011 } 8012 private void maybeForceBuilderInitialization() { 8013 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8014 } 8015 } 8016 private static Builder create() { 8017 return new Builder(); 8018 } 8019 8020 public Builder clear() { 8021 super.clear(); 8022 return this; 8023 } 8024 8025 public Builder clone() { 8026 return create().mergeFrom(buildPartial()); 8027 } 8028 8029 public com.google.protobuf.Descriptors.Descriptor 8030 getDescriptorForType() { 8031 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor; 8032 } 8033 8034 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto getDefaultInstanceForType() { 8035 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(); 8036 } 8037 8038 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto build() { 8039 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = buildPartial(); 8040 if (!result.isInitialized()) { 8041 throw newUninitializedMessageException(result); 8042 } 8043 return result; 8044 } 8045 8046 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto buildPartial() { 8047 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto(this); 8048 onBuilt(); 8049 return result; 8050 } 8051 8052 public Builder mergeFrom(com.google.protobuf.Message other) { 8053 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) { 8054 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)other); 8055 } else { 8056 super.mergeFrom(other); 8057 return this; 8058 } 8059 } 8060 8061 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other) { 8062 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()) return this; 8063 this.mergeUnknownFields(other.getUnknownFields()); 8064 return this; 8065 } 8066 8067 public final boolean isInitialized() { 8068 return true; 8069 } 8070 8071 public Builder mergeFrom( 8072 com.google.protobuf.CodedInputStream input, 8073 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8074 throws java.io.IOException { 8075 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parsedMessage = null; 8076 try { 8077 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 8078 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8079 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) e.getUnfinishedMessage(); 8080 throw e; 8081 } finally { 8082 if (parsedMessage != null) { 8083 mergeFrom(parsedMessage); 8084 } 8085 } 8086 return this; 8087 } 8088 8089 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsResponseProto) 8090 } 8091 8092 static { 8093 defaultInstance = new PurgeLogsResponseProto(true); 8094 defaultInstance.initFields(); 8095 } 8096 8097 // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsResponseProto) 8098 } 8099 8100 public interface IsFormattedRequestProtoOrBuilder 8101 extends com.google.protobuf.MessageOrBuilder { 8102 8103 // required .hadoop.hdfs.JournalIdProto jid = 1; 8104 /** 8105 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8106 */ 8107 boolean hasJid(); 8108 /** 8109 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8110 */ 8111 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid(); 8112 /** 8113 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8114 */ 8115 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder(); 8116 } 8117 /** 8118 * Protobuf type {@code hadoop.hdfs.IsFormattedRequestProto} 8119 * 8120 * <pre> 8121 ** 8122 * isFormatted() 8123 * </pre> 8124 */ 8125 public static final class IsFormattedRequestProto extends 8126 com.google.protobuf.GeneratedMessage 8127 implements IsFormattedRequestProtoOrBuilder { 8128 // Use IsFormattedRequestProto.newBuilder() to construct. 8129 private IsFormattedRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8130 super(builder); 8131 this.unknownFields = builder.getUnknownFields(); 8132 } 8133 private IsFormattedRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8134 8135 private static final IsFormattedRequestProto defaultInstance; 8136 public static IsFormattedRequestProto getDefaultInstance() { 8137 return defaultInstance; 8138 } 8139 8140 public IsFormattedRequestProto getDefaultInstanceForType() { 8141 return defaultInstance; 8142 } 8143 8144 private final com.google.protobuf.UnknownFieldSet unknownFields; 8145 @java.lang.Override 8146 public final com.google.protobuf.UnknownFieldSet 8147 getUnknownFields() { 8148 return this.unknownFields; 8149 } 8150 private IsFormattedRequestProto( 8151 com.google.protobuf.CodedInputStream input, 8152 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8153 throws com.google.protobuf.InvalidProtocolBufferException { 8154 initFields(); 8155 int mutable_bitField0_ = 0; 8156 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8157 com.google.protobuf.UnknownFieldSet.newBuilder(); 8158 try { 8159 boolean done = false; 8160 while (!done) { 8161 int tag = input.readTag(); 8162 switch (tag) { 8163 case 0: 8164 done = true; 8165 break; 8166 default: { 8167 if (!parseUnknownField(input, unknownFields, 8168 extensionRegistry, tag)) { 8169 done = true; 8170 } 8171 break; 8172 } 8173 case 10: { 8174 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 8175 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8176 subBuilder = jid_.toBuilder(); 8177 } 8178 jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 8179 if (subBuilder != null) { 8180 subBuilder.mergeFrom(jid_); 8181 jid_ = subBuilder.buildPartial(); 8182 } 8183 bitField0_ |= 0x00000001; 8184 break; 8185 } 8186 } 8187 } 8188 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8189 throw e.setUnfinishedMessage(this); 8190 } catch (java.io.IOException e) { 8191 throw new com.google.protobuf.InvalidProtocolBufferException( 8192 e.getMessage()).setUnfinishedMessage(this); 8193 } finally { 8194 this.unknownFields = unknownFields.build(); 8195 makeExtensionsImmutable(); 8196 } 8197 } 8198 public static final com.google.protobuf.Descriptors.Descriptor 8199 getDescriptor() { 8200 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor; 8201 } 8202 8203 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8204 internalGetFieldAccessorTable() { 8205 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable 8206 .ensureFieldAccessorsInitialized( 8207 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class); 8208 } 8209 8210 public static com.google.protobuf.Parser<IsFormattedRequestProto> PARSER = 8211 new com.google.protobuf.AbstractParser<IsFormattedRequestProto>() { 8212 public IsFormattedRequestProto parsePartialFrom( 8213 com.google.protobuf.CodedInputStream input, 8214 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8215 throws com.google.protobuf.InvalidProtocolBufferException { 8216 return new IsFormattedRequestProto(input, extensionRegistry); 8217 } 8218 }; 8219 8220 @java.lang.Override 8221 public com.google.protobuf.Parser<IsFormattedRequestProto> getParserForType() { 8222 return PARSER; 8223 } 8224 8225 private int bitField0_; 8226 // required .hadoop.hdfs.JournalIdProto jid = 1; 8227 public static final int JID_FIELD_NUMBER = 1; 8228 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_; 8229 /** 8230 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8231 */ 8232 public boolean hasJid() { 8233 return ((bitField0_ & 0x00000001) == 0x00000001); 8234 } 8235 /** 8236 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8237 */ 8238 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 8239 return jid_; 8240 } 8241 /** 8242 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8243 */ 8244 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 8245 return jid_; 8246 } 8247 8248 private void initFields() { 8249 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 8250 } 8251 private byte memoizedIsInitialized = -1; 8252 public final boolean isInitialized() { 8253 byte isInitialized = memoizedIsInitialized; 8254 if (isInitialized != -1) return isInitialized == 1; 8255 8256 if (!hasJid()) { 8257 memoizedIsInitialized = 0; 8258 return false; 8259 } 8260 if (!getJid().isInitialized()) { 8261 memoizedIsInitialized = 0; 8262 return false; 8263 } 8264 memoizedIsInitialized = 1; 8265 return true; 8266 } 8267 8268 public void writeTo(com.google.protobuf.CodedOutputStream output) 8269 throws java.io.IOException { 8270 getSerializedSize(); 8271 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8272 output.writeMessage(1, jid_); 8273 } 8274 getUnknownFields().writeTo(output); 8275 } 8276 8277 private int memoizedSerializedSize = -1; 8278 public int getSerializedSize() { 8279 int size = memoizedSerializedSize; 8280 if (size != -1) return size; 8281 8282 size = 0; 8283 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8284 size += com.google.protobuf.CodedOutputStream 8285 .computeMessageSize(1, jid_); 8286 } 8287 size += getUnknownFields().getSerializedSize(); 8288 memoizedSerializedSize = size; 8289 return size; 8290 } 8291 8292 private static final long serialVersionUID = 0L; 8293 @java.lang.Override 8294 protected java.lang.Object writeReplace() 8295 throws java.io.ObjectStreamException { 8296 return super.writeReplace(); 8297 } 8298 8299 @java.lang.Override 8300 public boolean equals(final java.lang.Object obj) { 8301 if (obj == this) { 8302 return true; 8303 } 8304 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)) { 8305 return super.equals(obj); 8306 } 8307 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) obj; 8308 8309 boolean result = true; 8310 result = result && (hasJid() == other.hasJid()); 8311 if (hasJid()) { 8312 result = result && getJid() 8313 .equals(other.getJid()); 8314 } 8315 result = result && 8316 getUnknownFields().equals(other.getUnknownFields()); 8317 return result; 8318 } 8319 8320 private int memoizedHashCode = 0; 8321 @java.lang.Override 8322 public int hashCode() { 8323 if (memoizedHashCode != 0) { 8324 return memoizedHashCode; 8325 } 8326 int hash = 41; 8327 hash = (19 * hash) + getDescriptorForType().hashCode(); 8328 if (hasJid()) { 8329 hash = (37 * hash) + JID_FIELD_NUMBER; 8330 hash = (53 * hash) + getJid().hashCode(); 8331 } 8332 hash = (29 * hash) + getUnknownFields().hashCode(); 8333 memoizedHashCode = hash; 8334 return hash; 8335 } 8336 8337 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8338 com.google.protobuf.ByteString data) 8339 throws com.google.protobuf.InvalidProtocolBufferException { 8340 return PARSER.parseFrom(data); 8341 } 8342 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8343 com.google.protobuf.ByteString data, 8344 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8345 throws com.google.protobuf.InvalidProtocolBufferException { 8346 return PARSER.parseFrom(data, extensionRegistry); 8347 } 8348 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(byte[] data) 8349 throws com.google.protobuf.InvalidProtocolBufferException { 8350 return PARSER.parseFrom(data); 8351 } 8352 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8353 byte[] data, 8354 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8355 throws com.google.protobuf.InvalidProtocolBufferException { 8356 return PARSER.parseFrom(data, extensionRegistry); 8357 } 8358 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(java.io.InputStream input) 8359 throws java.io.IOException { 8360 return PARSER.parseFrom(input); 8361 } 8362 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8363 java.io.InputStream input, 8364 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8365 throws java.io.IOException { 8366 return PARSER.parseFrom(input, extensionRegistry); 8367 } 8368 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(java.io.InputStream input) 8369 throws java.io.IOException { 8370 return PARSER.parseDelimitedFrom(input); 8371 } 8372 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom( 8373 java.io.InputStream input, 8374 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8375 throws java.io.IOException { 8376 return PARSER.parseDelimitedFrom(input, extensionRegistry); 8377 } 8378 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8379 com.google.protobuf.CodedInputStream input) 8380 throws java.io.IOException { 8381 return PARSER.parseFrom(input); 8382 } 8383 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom( 8384 com.google.protobuf.CodedInputStream input, 8385 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8386 throws java.io.IOException { 8387 return PARSER.parseFrom(input, extensionRegistry); 8388 } 8389 8390 public static Builder newBuilder() { return Builder.create(); } 8391 public Builder newBuilderForType() { return newBuilder(); } 8392 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto prototype) { 8393 return newBuilder().mergeFrom(prototype); 8394 } 8395 public Builder toBuilder() { return newBuilder(this); } 8396 8397 @java.lang.Override 8398 protected Builder newBuilderForType( 8399 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8400 Builder builder = new Builder(parent); 8401 return builder; 8402 } 8403 /** 8404 * Protobuf type {@code hadoop.hdfs.IsFormattedRequestProto} 8405 * 8406 * <pre> 8407 ** 8408 * isFormatted() 8409 * </pre> 8410 */ 8411 public static final class Builder extends 8412 com.google.protobuf.GeneratedMessage.Builder<Builder> 8413 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProtoOrBuilder { 8414 public static final com.google.protobuf.Descriptors.Descriptor 8415 getDescriptor() { 8416 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor; 8417 } 8418 8419 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8420 internalGetFieldAccessorTable() { 8421 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable 8422 .ensureFieldAccessorsInitialized( 8423 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class); 8424 } 8425 8426 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.newBuilder() 8427 private Builder() { 8428 maybeForceBuilderInitialization(); 8429 } 8430 8431 private Builder( 8432 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8433 super(parent); 8434 maybeForceBuilderInitialization(); 8435 } 8436 private void maybeForceBuilderInitialization() { 8437 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8438 getJidFieldBuilder(); 8439 } 8440 } 8441 private static Builder create() { 8442 return new Builder(); 8443 } 8444 8445 public Builder clear() { 8446 super.clear(); 8447 if (jidBuilder_ == null) { 8448 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 8449 } else { 8450 jidBuilder_.clear(); 8451 } 8452 bitField0_ = (bitField0_ & ~0x00000001); 8453 return this; 8454 } 8455 8456 public Builder clone() { 8457 return create().mergeFrom(buildPartial()); 8458 } 8459 8460 public com.google.protobuf.Descriptors.Descriptor 8461 getDescriptorForType() { 8462 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor; 8463 } 8464 8465 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto getDefaultInstanceForType() { 8466 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance(); 8467 } 8468 8469 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto build() { 8470 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = buildPartial(); 8471 if (!result.isInitialized()) { 8472 throw newUninitializedMessageException(result); 8473 } 8474 return result; 8475 } 8476 8477 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto buildPartial() { 8478 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto(this); 8479 int from_bitField0_ = bitField0_; 8480 int to_bitField0_ = 0; 8481 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 8482 to_bitField0_ |= 0x00000001; 8483 } 8484 if (jidBuilder_ == null) { 8485 result.jid_ = jid_; 8486 } else { 8487 result.jid_ = jidBuilder_.build(); 8488 } 8489 result.bitField0_ = to_bitField0_; 8490 onBuilt(); 8491 return result; 8492 } 8493 8494 public Builder mergeFrom(com.google.protobuf.Message other) { 8495 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) { 8496 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)other); 8497 } else { 8498 super.mergeFrom(other); 8499 return this; 8500 } 8501 } 8502 8503 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other) { 8504 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance()) return this; 8505 if (other.hasJid()) { 8506 mergeJid(other.getJid()); 8507 } 8508 this.mergeUnknownFields(other.getUnknownFields()); 8509 return this; 8510 } 8511 8512 public final boolean isInitialized() { 8513 if (!hasJid()) { 8514 8515 return false; 8516 } 8517 if (!getJid().isInitialized()) { 8518 8519 return false; 8520 } 8521 return true; 8522 } 8523 8524 public Builder mergeFrom( 8525 com.google.protobuf.CodedInputStream input, 8526 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8527 throws java.io.IOException { 8528 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parsedMessage = null; 8529 try { 8530 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 8531 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8532 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) e.getUnfinishedMessage(); 8533 throw e; 8534 } finally { 8535 if (parsedMessage != null) { 8536 mergeFrom(parsedMessage); 8537 } 8538 } 8539 return this; 8540 } 8541 private int bitField0_; 8542 8543 // required .hadoop.hdfs.JournalIdProto jid = 1; 8544 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 8545 private com.google.protobuf.SingleFieldBuilder< 8546 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_; 8547 /** 8548 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8549 */ 8550 public boolean hasJid() { 8551 return ((bitField0_ & 0x00000001) == 0x00000001); 8552 } 8553 /** 8554 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8555 */ 8556 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 8557 if (jidBuilder_ == null) { 8558 return jid_; 8559 } else { 8560 return jidBuilder_.getMessage(); 8561 } 8562 } 8563 /** 8564 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8565 */ 8566 public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 8567 if (jidBuilder_ == null) { 8568 if (value == null) { 8569 throw new NullPointerException(); 8570 } 8571 jid_ = value; 8572 onChanged(); 8573 } else { 8574 jidBuilder_.setMessage(value); 8575 } 8576 bitField0_ |= 0x00000001; 8577 return this; 8578 } 8579 /** 8580 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8581 */ 8582 public Builder setJid( 8583 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 8584 if (jidBuilder_ == null) { 8585 jid_ = builderForValue.build(); 8586 onChanged(); 8587 } else { 8588 jidBuilder_.setMessage(builderForValue.build()); 8589 } 8590 bitField0_ |= 0x00000001; 8591 return this; 8592 } 8593 /** 8594 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8595 */ 8596 public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 8597 if (jidBuilder_ == null) { 8598 if (((bitField0_ & 0x00000001) == 0x00000001) && 8599 jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 8600 jid_ = 8601 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial(); 8602 } else { 8603 jid_ = value; 8604 } 8605 onChanged(); 8606 } else { 8607 jidBuilder_.mergeFrom(value); 8608 } 8609 bitField0_ |= 0x00000001; 8610 return this; 8611 } 8612 /** 8613 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8614 */ 8615 public Builder clearJid() { 8616 if (jidBuilder_ == null) { 8617 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 8618 onChanged(); 8619 } else { 8620 jidBuilder_.clear(); 8621 } 8622 bitField0_ = (bitField0_ & ~0x00000001); 8623 return this; 8624 } 8625 /** 8626 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8627 */ 8628 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() { 8629 bitField0_ |= 0x00000001; 8630 onChanged(); 8631 return getJidFieldBuilder().getBuilder(); 8632 } 8633 /** 8634 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8635 */ 8636 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 8637 if (jidBuilder_ != null) { 8638 return jidBuilder_.getMessageOrBuilder(); 8639 } else { 8640 return jid_; 8641 } 8642 } 8643 /** 8644 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 8645 */ 8646 private com.google.protobuf.SingleFieldBuilder< 8647 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 8648 getJidFieldBuilder() { 8649 if (jidBuilder_ == null) { 8650 jidBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8651 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 8652 jid_, 8653 getParentForChildren(), 8654 isClean()); 8655 jid_ = null; 8656 } 8657 return jidBuilder_; 8658 } 8659 8660 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedRequestProto) 8661 } 8662 8663 static { 8664 defaultInstance = new IsFormattedRequestProto(true); 8665 defaultInstance.initFields(); 8666 } 8667 8668 // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedRequestProto) 8669 } 8670 8671 public interface IsFormattedResponseProtoOrBuilder 8672 extends com.google.protobuf.MessageOrBuilder { 8673 8674 // required bool isFormatted = 1; 8675 /** 8676 * <code>required bool isFormatted = 1;</code> 8677 */ 8678 boolean hasIsFormatted(); 8679 /** 8680 * <code>required bool isFormatted = 1;</code> 8681 */ 8682 boolean getIsFormatted(); 8683 } 8684 /** 8685 * Protobuf type {@code hadoop.hdfs.IsFormattedResponseProto} 8686 */ 8687 public static final class IsFormattedResponseProto extends 8688 com.google.protobuf.GeneratedMessage 8689 implements IsFormattedResponseProtoOrBuilder { 8690 // Use IsFormattedResponseProto.newBuilder() to construct. 8691 private IsFormattedResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8692 super(builder); 8693 this.unknownFields = builder.getUnknownFields(); 8694 } 8695 private IsFormattedResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8696 8697 private static final IsFormattedResponseProto defaultInstance; 8698 public static IsFormattedResponseProto getDefaultInstance() { 8699 return defaultInstance; 8700 } 8701 8702 public IsFormattedResponseProto getDefaultInstanceForType() { 8703 return defaultInstance; 8704 } 8705 8706 private final com.google.protobuf.UnknownFieldSet unknownFields; 8707 @java.lang.Override 8708 public final com.google.protobuf.UnknownFieldSet 8709 getUnknownFields() { 8710 return this.unknownFields; 8711 } 8712 private IsFormattedResponseProto( 8713 com.google.protobuf.CodedInputStream input, 8714 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8715 throws com.google.protobuf.InvalidProtocolBufferException { 8716 initFields(); 8717 int mutable_bitField0_ = 0; 8718 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8719 com.google.protobuf.UnknownFieldSet.newBuilder(); 8720 try { 8721 boolean done = false; 8722 while (!done) { 8723 int tag = input.readTag(); 8724 switch (tag) { 8725 case 0: 8726 done = true; 8727 break; 8728 default: { 8729 if (!parseUnknownField(input, unknownFields, 8730 extensionRegistry, tag)) { 8731 done = true; 8732 } 8733 break; 8734 } 8735 case 8: { 8736 bitField0_ |= 0x00000001; 8737 isFormatted_ = input.readBool(); 8738 break; 8739 } 8740 } 8741 } 8742 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8743 throw e.setUnfinishedMessage(this); 8744 } catch (java.io.IOException e) { 8745 throw new com.google.protobuf.InvalidProtocolBufferException( 8746 e.getMessage()).setUnfinishedMessage(this); 8747 } finally { 8748 this.unknownFields = unknownFields.build(); 8749 makeExtensionsImmutable(); 8750 } 8751 } 8752 public static final com.google.protobuf.Descriptors.Descriptor 8753 getDescriptor() { 8754 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor; 8755 } 8756 8757 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8758 internalGetFieldAccessorTable() { 8759 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable 8760 .ensureFieldAccessorsInitialized( 8761 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class); 8762 } 8763 8764 public static com.google.protobuf.Parser<IsFormattedResponseProto> PARSER = 8765 new com.google.protobuf.AbstractParser<IsFormattedResponseProto>() { 8766 public IsFormattedResponseProto parsePartialFrom( 8767 com.google.protobuf.CodedInputStream input, 8768 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8769 throws com.google.protobuf.InvalidProtocolBufferException { 8770 return new IsFormattedResponseProto(input, extensionRegistry); 8771 } 8772 }; 8773 8774 @java.lang.Override 8775 public com.google.protobuf.Parser<IsFormattedResponseProto> getParserForType() { 8776 return PARSER; 8777 } 8778 8779 private int bitField0_; 8780 // required bool isFormatted = 1; 8781 public static final int ISFORMATTED_FIELD_NUMBER = 1; 8782 private boolean isFormatted_; 8783 /** 8784 * <code>required bool isFormatted = 1;</code> 8785 */ 8786 public boolean hasIsFormatted() { 8787 return ((bitField0_ & 0x00000001) == 0x00000001); 8788 } 8789 /** 8790 * <code>required bool isFormatted = 1;</code> 8791 */ 8792 public boolean getIsFormatted() { 8793 return isFormatted_; 8794 } 8795 8796 private void initFields() { 8797 isFormatted_ = false; 8798 } 8799 private byte memoizedIsInitialized = -1; 8800 public final boolean isInitialized() { 8801 byte isInitialized = memoizedIsInitialized; 8802 if (isInitialized != -1) return isInitialized == 1; 8803 8804 if (!hasIsFormatted()) { 8805 memoizedIsInitialized = 0; 8806 return false; 8807 } 8808 memoizedIsInitialized = 1; 8809 return true; 8810 } 8811 8812 public void writeTo(com.google.protobuf.CodedOutputStream output) 8813 throws java.io.IOException { 8814 getSerializedSize(); 8815 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8816 output.writeBool(1, isFormatted_); 8817 } 8818 getUnknownFields().writeTo(output); 8819 } 8820 8821 private int memoizedSerializedSize = -1; 8822 public int getSerializedSize() { 8823 int size = memoizedSerializedSize; 8824 if (size != -1) return size; 8825 8826 size = 0; 8827 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8828 size += com.google.protobuf.CodedOutputStream 8829 .computeBoolSize(1, isFormatted_); 8830 } 8831 size += getUnknownFields().getSerializedSize(); 8832 memoizedSerializedSize = size; 8833 return size; 8834 } 8835 8836 private static final long serialVersionUID = 0L; 8837 @java.lang.Override 8838 protected java.lang.Object writeReplace() 8839 throws java.io.ObjectStreamException { 8840 return super.writeReplace(); 8841 } 8842 8843 @java.lang.Override 8844 public boolean equals(final java.lang.Object obj) { 8845 if (obj == this) { 8846 return true; 8847 } 8848 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)) { 8849 return super.equals(obj); 8850 } 8851 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) obj; 8852 8853 boolean result = true; 8854 result = result && (hasIsFormatted() == other.hasIsFormatted()); 8855 if (hasIsFormatted()) { 8856 result = result && (getIsFormatted() 8857 == other.getIsFormatted()); 8858 } 8859 result = result && 8860 getUnknownFields().equals(other.getUnknownFields()); 8861 return result; 8862 } 8863 8864 private int memoizedHashCode = 0; 8865 @java.lang.Override 8866 public int hashCode() { 8867 if (memoizedHashCode != 0) { 8868 return memoizedHashCode; 8869 } 8870 int hash = 41; 8871 hash = (19 * hash) + getDescriptorForType().hashCode(); 8872 if (hasIsFormatted()) { 8873 hash = (37 * hash) + ISFORMATTED_FIELD_NUMBER; 8874 hash = (53 * hash) + hashBoolean(getIsFormatted()); 8875 } 8876 hash = (29 * hash) + getUnknownFields().hashCode(); 8877 memoizedHashCode = hash; 8878 return hash; 8879 } 8880 8881 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8882 com.google.protobuf.ByteString data) 8883 throws com.google.protobuf.InvalidProtocolBufferException { 8884 return PARSER.parseFrom(data); 8885 } 8886 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8887 com.google.protobuf.ByteString data, 8888 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8889 throws com.google.protobuf.InvalidProtocolBufferException { 8890 return PARSER.parseFrom(data, extensionRegistry); 8891 } 8892 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(byte[] data) 8893 throws com.google.protobuf.InvalidProtocolBufferException { 8894 return PARSER.parseFrom(data); 8895 } 8896 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8897 byte[] data, 8898 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8899 throws com.google.protobuf.InvalidProtocolBufferException { 8900 return PARSER.parseFrom(data, extensionRegistry); 8901 } 8902 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(java.io.InputStream input) 8903 throws java.io.IOException { 8904 return PARSER.parseFrom(input); 8905 } 8906 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8907 java.io.InputStream input, 8908 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8909 throws java.io.IOException { 8910 return PARSER.parseFrom(input, extensionRegistry); 8911 } 8912 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(java.io.InputStream input) 8913 throws java.io.IOException { 8914 return PARSER.parseDelimitedFrom(input); 8915 } 8916 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom( 8917 java.io.InputStream input, 8918 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8919 throws java.io.IOException { 8920 return PARSER.parseDelimitedFrom(input, extensionRegistry); 8921 } 8922 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8923 com.google.protobuf.CodedInputStream input) 8924 throws java.io.IOException { 8925 return PARSER.parseFrom(input); 8926 } 8927 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom( 8928 com.google.protobuf.CodedInputStream input, 8929 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8930 throws java.io.IOException { 8931 return PARSER.parseFrom(input, extensionRegistry); 8932 } 8933 8934 public static Builder newBuilder() { return Builder.create(); } 8935 public Builder newBuilderForType() { return newBuilder(); } 8936 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto prototype) { 8937 return newBuilder().mergeFrom(prototype); 8938 } 8939 public Builder toBuilder() { return newBuilder(this); } 8940 8941 @java.lang.Override 8942 protected Builder newBuilderForType( 8943 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8944 Builder builder = new Builder(parent); 8945 return builder; 8946 } 8947 /** 8948 * Protobuf type {@code hadoop.hdfs.IsFormattedResponseProto} 8949 */ 8950 public static final class Builder extends 8951 com.google.protobuf.GeneratedMessage.Builder<Builder> 8952 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProtoOrBuilder { 8953 public static final com.google.protobuf.Descriptors.Descriptor 8954 getDescriptor() { 8955 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor; 8956 } 8957 8958 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8959 internalGetFieldAccessorTable() { 8960 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable 8961 .ensureFieldAccessorsInitialized( 8962 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class); 8963 } 8964 8965 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.newBuilder() 8966 private Builder() { 8967 maybeForceBuilderInitialization(); 8968 } 8969 8970 private Builder( 8971 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8972 super(parent); 8973 maybeForceBuilderInitialization(); 8974 } 8975 private void maybeForceBuilderInitialization() { 8976 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8977 } 8978 } 8979 private static Builder create() { 8980 return new Builder(); 8981 } 8982 8983 public Builder clear() { 8984 super.clear(); 8985 isFormatted_ = false; 8986 bitField0_ = (bitField0_ & ~0x00000001); 8987 return this; 8988 } 8989 8990 public Builder clone() { 8991 return create().mergeFrom(buildPartial()); 8992 } 8993 8994 public com.google.protobuf.Descriptors.Descriptor 8995 getDescriptorForType() { 8996 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor; 8997 } 8998 8999 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto getDefaultInstanceForType() { 9000 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(); 9001 } 9002 9003 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto build() { 9004 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = buildPartial(); 9005 if (!result.isInitialized()) { 9006 throw newUninitializedMessageException(result); 9007 } 9008 return result; 9009 } 9010 9011 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto buildPartial() { 9012 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto(this); 9013 int from_bitField0_ = bitField0_; 9014 int to_bitField0_ = 0; 9015 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 9016 to_bitField0_ |= 0x00000001; 9017 } 9018 result.isFormatted_ = isFormatted_; 9019 result.bitField0_ = to_bitField0_; 9020 onBuilt(); 9021 return result; 9022 } 9023 9024 public Builder mergeFrom(com.google.protobuf.Message other) { 9025 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) { 9026 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)other); 9027 } else { 9028 super.mergeFrom(other); 9029 return this; 9030 } 9031 } 9032 9033 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other) { 9034 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()) return this; 9035 if (other.hasIsFormatted()) { 9036 setIsFormatted(other.getIsFormatted()); 9037 } 9038 this.mergeUnknownFields(other.getUnknownFields()); 9039 return this; 9040 } 9041 9042 public final boolean isInitialized() { 9043 if (!hasIsFormatted()) { 9044 9045 return false; 9046 } 9047 return true; 9048 } 9049 9050 public Builder mergeFrom( 9051 com.google.protobuf.CodedInputStream input, 9052 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9053 throws java.io.IOException { 9054 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parsedMessage = null; 9055 try { 9056 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 9057 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9058 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) e.getUnfinishedMessage(); 9059 throw e; 9060 } finally { 9061 if (parsedMessage != null) { 9062 mergeFrom(parsedMessage); 9063 } 9064 } 9065 return this; 9066 } 9067 private int bitField0_; 9068 9069 // required bool isFormatted = 1; 9070 private boolean isFormatted_ ; 9071 /** 9072 * <code>required bool isFormatted = 1;</code> 9073 */ 9074 public boolean hasIsFormatted() { 9075 return ((bitField0_ & 0x00000001) == 0x00000001); 9076 } 9077 /** 9078 * <code>required bool isFormatted = 1;</code> 9079 */ 9080 public boolean getIsFormatted() { 9081 return isFormatted_; 9082 } 9083 /** 9084 * <code>required bool isFormatted = 1;</code> 9085 */ 9086 public Builder setIsFormatted(boolean value) { 9087 bitField0_ |= 0x00000001; 9088 isFormatted_ = value; 9089 onChanged(); 9090 return this; 9091 } 9092 /** 9093 * <code>required bool isFormatted = 1;</code> 9094 */ 9095 public Builder clearIsFormatted() { 9096 bitField0_ = (bitField0_ & ~0x00000001); 9097 isFormatted_ = false; 9098 onChanged(); 9099 return this; 9100 } 9101 9102 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedResponseProto) 9103 } 9104 9105 static { 9106 defaultInstance = new IsFormattedResponseProto(true); 9107 defaultInstance.initFields(); 9108 } 9109 9110 // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedResponseProto) 9111 } 9112 9113 public interface GetJournalStateRequestProtoOrBuilder 9114 extends com.google.protobuf.MessageOrBuilder { 9115 9116 // required .hadoop.hdfs.JournalIdProto jid = 1; 9117 /** 9118 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9119 */ 9120 boolean hasJid(); 9121 /** 9122 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9123 */ 9124 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid(); 9125 /** 9126 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9127 */ 9128 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder(); 9129 } 9130 /** 9131 * Protobuf type {@code hadoop.hdfs.GetJournalStateRequestProto} 9132 * 9133 * <pre> 9134 ** 9135 * getJournalState() 9136 * </pre> 9137 */ 9138 public static final class GetJournalStateRequestProto extends 9139 com.google.protobuf.GeneratedMessage 9140 implements GetJournalStateRequestProtoOrBuilder { 9141 // Use GetJournalStateRequestProto.newBuilder() to construct. 9142 private GetJournalStateRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 9143 super(builder); 9144 this.unknownFields = builder.getUnknownFields(); 9145 } 9146 private GetJournalStateRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 9147 9148 private static final GetJournalStateRequestProto defaultInstance; 9149 public static GetJournalStateRequestProto getDefaultInstance() { 9150 return defaultInstance; 9151 } 9152 9153 public GetJournalStateRequestProto getDefaultInstanceForType() { 9154 return defaultInstance; 9155 } 9156 9157 private final com.google.protobuf.UnknownFieldSet unknownFields; 9158 @java.lang.Override 9159 public final com.google.protobuf.UnknownFieldSet 9160 getUnknownFields() { 9161 return this.unknownFields; 9162 } 9163 private GetJournalStateRequestProto( 9164 com.google.protobuf.CodedInputStream input, 9165 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9166 throws com.google.protobuf.InvalidProtocolBufferException { 9167 initFields(); 9168 int mutable_bitField0_ = 0; 9169 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 9170 com.google.protobuf.UnknownFieldSet.newBuilder(); 9171 try { 9172 boolean done = false; 9173 while (!done) { 9174 int tag = input.readTag(); 9175 switch (tag) { 9176 case 0: 9177 done = true; 9178 break; 9179 default: { 9180 if (!parseUnknownField(input, unknownFields, 9181 extensionRegistry, tag)) { 9182 done = true; 9183 } 9184 break; 9185 } 9186 case 10: { 9187 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 9188 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9189 subBuilder = jid_.toBuilder(); 9190 } 9191 jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 9192 if (subBuilder != null) { 9193 subBuilder.mergeFrom(jid_); 9194 jid_ = subBuilder.buildPartial(); 9195 } 9196 bitField0_ |= 0x00000001; 9197 break; 9198 } 9199 } 9200 } 9201 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9202 throw e.setUnfinishedMessage(this); 9203 } catch (java.io.IOException e) { 9204 throw new com.google.protobuf.InvalidProtocolBufferException( 9205 e.getMessage()).setUnfinishedMessage(this); 9206 } finally { 9207 this.unknownFields = unknownFields.build(); 9208 makeExtensionsImmutable(); 9209 } 9210 } 9211 public static final com.google.protobuf.Descriptors.Descriptor 9212 getDescriptor() { 9213 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor; 9214 } 9215 9216 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9217 internalGetFieldAccessorTable() { 9218 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable 9219 .ensureFieldAccessorsInitialized( 9220 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class); 9221 } 9222 9223 public static com.google.protobuf.Parser<GetJournalStateRequestProto> PARSER = 9224 new com.google.protobuf.AbstractParser<GetJournalStateRequestProto>() { 9225 public GetJournalStateRequestProto parsePartialFrom( 9226 com.google.protobuf.CodedInputStream input, 9227 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9228 throws com.google.protobuf.InvalidProtocolBufferException { 9229 return new GetJournalStateRequestProto(input, extensionRegistry); 9230 } 9231 }; 9232 9233 @java.lang.Override 9234 public com.google.protobuf.Parser<GetJournalStateRequestProto> getParserForType() { 9235 return PARSER; 9236 } 9237 9238 private int bitField0_; 9239 // required .hadoop.hdfs.JournalIdProto jid = 1; 9240 public static final int JID_FIELD_NUMBER = 1; 9241 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_; 9242 /** 9243 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9244 */ 9245 public boolean hasJid() { 9246 return ((bitField0_ & 0x00000001) == 0x00000001); 9247 } 9248 /** 9249 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9250 */ 9251 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 9252 return jid_; 9253 } 9254 /** 9255 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9256 */ 9257 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 9258 return jid_; 9259 } 9260 9261 private void initFields() { 9262 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 9263 } 9264 private byte memoizedIsInitialized = -1; 9265 public final boolean isInitialized() { 9266 byte isInitialized = memoizedIsInitialized; 9267 if (isInitialized != -1) return isInitialized == 1; 9268 9269 if (!hasJid()) { 9270 memoizedIsInitialized = 0; 9271 return false; 9272 } 9273 if (!getJid().isInitialized()) { 9274 memoizedIsInitialized = 0; 9275 return false; 9276 } 9277 memoizedIsInitialized = 1; 9278 return true; 9279 } 9280 9281 public void writeTo(com.google.protobuf.CodedOutputStream output) 9282 throws java.io.IOException { 9283 getSerializedSize(); 9284 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9285 output.writeMessage(1, jid_); 9286 } 9287 getUnknownFields().writeTo(output); 9288 } 9289 9290 private int memoizedSerializedSize = -1; 9291 public int getSerializedSize() { 9292 int size = memoizedSerializedSize; 9293 if (size != -1) return size; 9294 9295 size = 0; 9296 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9297 size += com.google.protobuf.CodedOutputStream 9298 .computeMessageSize(1, jid_); 9299 } 9300 size += getUnknownFields().getSerializedSize(); 9301 memoizedSerializedSize = size; 9302 return size; 9303 } 9304 9305 private static final long serialVersionUID = 0L; 9306 @java.lang.Override 9307 protected java.lang.Object writeReplace() 9308 throws java.io.ObjectStreamException { 9309 return super.writeReplace(); 9310 } 9311 9312 @java.lang.Override 9313 public boolean equals(final java.lang.Object obj) { 9314 if (obj == this) { 9315 return true; 9316 } 9317 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)) { 9318 return super.equals(obj); 9319 } 9320 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) obj; 9321 9322 boolean result = true; 9323 result = result && (hasJid() == other.hasJid()); 9324 if (hasJid()) { 9325 result = result && getJid() 9326 .equals(other.getJid()); 9327 } 9328 result = result && 9329 getUnknownFields().equals(other.getUnknownFields()); 9330 return result; 9331 } 9332 9333 private int memoizedHashCode = 0; 9334 @java.lang.Override 9335 public int hashCode() { 9336 if (memoizedHashCode != 0) { 9337 return memoizedHashCode; 9338 } 9339 int hash = 41; 9340 hash = (19 * hash) + getDescriptorForType().hashCode(); 9341 if (hasJid()) { 9342 hash = (37 * hash) + JID_FIELD_NUMBER; 9343 hash = (53 * hash) + getJid().hashCode(); 9344 } 9345 hash = (29 * hash) + getUnknownFields().hashCode(); 9346 memoizedHashCode = hash; 9347 return hash; 9348 } 9349 9350 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9351 com.google.protobuf.ByteString data) 9352 throws com.google.protobuf.InvalidProtocolBufferException { 9353 return PARSER.parseFrom(data); 9354 } 9355 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9356 com.google.protobuf.ByteString data, 9357 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9358 throws com.google.protobuf.InvalidProtocolBufferException { 9359 return PARSER.parseFrom(data, extensionRegistry); 9360 } 9361 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(byte[] data) 9362 throws com.google.protobuf.InvalidProtocolBufferException { 9363 return PARSER.parseFrom(data); 9364 } 9365 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9366 byte[] data, 9367 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9368 throws com.google.protobuf.InvalidProtocolBufferException { 9369 return PARSER.parseFrom(data, extensionRegistry); 9370 } 9371 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(java.io.InputStream input) 9372 throws java.io.IOException { 9373 return PARSER.parseFrom(input); 9374 } 9375 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9376 java.io.InputStream input, 9377 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9378 throws java.io.IOException { 9379 return PARSER.parseFrom(input, extensionRegistry); 9380 } 9381 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(java.io.InputStream input) 9382 throws java.io.IOException { 9383 return PARSER.parseDelimitedFrom(input); 9384 } 9385 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom( 9386 java.io.InputStream input, 9387 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9388 throws java.io.IOException { 9389 return PARSER.parseDelimitedFrom(input, extensionRegistry); 9390 } 9391 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9392 com.google.protobuf.CodedInputStream input) 9393 throws java.io.IOException { 9394 return PARSER.parseFrom(input); 9395 } 9396 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom( 9397 com.google.protobuf.CodedInputStream input, 9398 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9399 throws java.io.IOException { 9400 return PARSER.parseFrom(input, extensionRegistry); 9401 } 9402 9403 public static Builder newBuilder() { return Builder.create(); } 9404 public Builder newBuilderForType() { return newBuilder(); } 9405 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto prototype) { 9406 return newBuilder().mergeFrom(prototype); 9407 } 9408 public Builder toBuilder() { return newBuilder(this); } 9409 9410 @java.lang.Override 9411 protected Builder newBuilderForType( 9412 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9413 Builder builder = new Builder(parent); 9414 return builder; 9415 } 9416 /** 9417 * Protobuf type {@code hadoop.hdfs.GetJournalStateRequestProto} 9418 * 9419 * <pre> 9420 ** 9421 * getJournalState() 9422 * </pre> 9423 */ 9424 public static final class Builder extends 9425 com.google.protobuf.GeneratedMessage.Builder<Builder> 9426 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProtoOrBuilder { 9427 public static final com.google.protobuf.Descriptors.Descriptor 9428 getDescriptor() { 9429 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor; 9430 } 9431 9432 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9433 internalGetFieldAccessorTable() { 9434 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable 9435 .ensureFieldAccessorsInitialized( 9436 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class); 9437 } 9438 9439 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.newBuilder() 9440 private Builder() { 9441 maybeForceBuilderInitialization(); 9442 } 9443 9444 private Builder( 9445 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9446 super(parent); 9447 maybeForceBuilderInitialization(); 9448 } 9449 private void maybeForceBuilderInitialization() { 9450 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 9451 getJidFieldBuilder(); 9452 } 9453 } 9454 private static Builder create() { 9455 return new Builder(); 9456 } 9457 9458 public Builder clear() { 9459 super.clear(); 9460 if (jidBuilder_ == null) { 9461 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 9462 } else { 9463 jidBuilder_.clear(); 9464 } 9465 bitField0_ = (bitField0_ & ~0x00000001); 9466 return this; 9467 } 9468 9469 public Builder clone() { 9470 return create().mergeFrom(buildPartial()); 9471 } 9472 9473 public com.google.protobuf.Descriptors.Descriptor 9474 getDescriptorForType() { 9475 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor; 9476 } 9477 9478 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto getDefaultInstanceForType() { 9479 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance(); 9480 } 9481 9482 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto build() { 9483 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = buildPartial(); 9484 if (!result.isInitialized()) { 9485 throw newUninitializedMessageException(result); 9486 } 9487 return result; 9488 } 9489 9490 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto buildPartial() { 9491 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto(this); 9492 int from_bitField0_ = bitField0_; 9493 int to_bitField0_ = 0; 9494 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 9495 to_bitField0_ |= 0x00000001; 9496 } 9497 if (jidBuilder_ == null) { 9498 result.jid_ = jid_; 9499 } else { 9500 result.jid_ = jidBuilder_.build(); 9501 } 9502 result.bitField0_ = to_bitField0_; 9503 onBuilt(); 9504 return result; 9505 } 9506 9507 public Builder mergeFrom(com.google.protobuf.Message other) { 9508 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) { 9509 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)other); 9510 } else { 9511 super.mergeFrom(other); 9512 return this; 9513 } 9514 } 9515 9516 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other) { 9517 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance()) return this; 9518 if (other.hasJid()) { 9519 mergeJid(other.getJid()); 9520 } 9521 this.mergeUnknownFields(other.getUnknownFields()); 9522 return this; 9523 } 9524 9525 public final boolean isInitialized() { 9526 if (!hasJid()) { 9527 9528 return false; 9529 } 9530 if (!getJid().isInitialized()) { 9531 9532 return false; 9533 } 9534 return true; 9535 } 9536 9537 public Builder mergeFrom( 9538 com.google.protobuf.CodedInputStream input, 9539 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9540 throws java.io.IOException { 9541 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parsedMessage = null; 9542 try { 9543 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 9544 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9545 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) e.getUnfinishedMessage(); 9546 throw e; 9547 } finally { 9548 if (parsedMessage != null) { 9549 mergeFrom(parsedMessage); 9550 } 9551 } 9552 return this; 9553 } 9554 private int bitField0_; 9555 9556 // required .hadoop.hdfs.JournalIdProto jid = 1; 9557 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 9558 private com.google.protobuf.SingleFieldBuilder< 9559 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_; 9560 /** 9561 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9562 */ 9563 public boolean hasJid() { 9564 return ((bitField0_ & 0x00000001) == 0x00000001); 9565 } 9566 /** 9567 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9568 */ 9569 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 9570 if (jidBuilder_ == null) { 9571 return jid_; 9572 } else { 9573 return jidBuilder_.getMessage(); 9574 } 9575 } 9576 /** 9577 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9578 */ 9579 public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 9580 if (jidBuilder_ == null) { 9581 if (value == null) { 9582 throw new NullPointerException(); 9583 } 9584 jid_ = value; 9585 onChanged(); 9586 } else { 9587 jidBuilder_.setMessage(value); 9588 } 9589 bitField0_ |= 0x00000001; 9590 return this; 9591 } 9592 /** 9593 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9594 */ 9595 public Builder setJid( 9596 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 9597 if (jidBuilder_ == null) { 9598 jid_ = builderForValue.build(); 9599 onChanged(); 9600 } else { 9601 jidBuilder_.setMessage(builderForValue.build()); 9602 } 9603 bitField0_ |= 0x00000001; 9604 return this; 9605 } 9606 /** 9607 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9608 */ 9609 public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 9610 if (jidBuilder_ == null) { 9611 if (((bitField0_ & 0x00000001) == 0x00000001) && 9612 jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 9613 jid_ = 9614 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial(); 9615 } else { 9616 jid_ = value; 9617 } 9618 onChanged(); 9619 } else { 9620 jidBuilder_.mergeFrom(value); 9621 } 9622 bitField0_ |= 0x00000001; 9623 return this; 9624 } 9625 /** 9626 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9627 */ 9628 public Builder clearJid() { 9629 if (jidBuilder_ == null) { 9630 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 9631 onChanged(); 9632 } else { 9633 jidBuilder_.clear(); 9634 } 9635 bitField0_ = (bitField0_ & ~0x00000001); 9636 return this; 9637 } 9638 /** 9639 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9640 */ 9641 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() { 9642 bitField0_ |= 0x00000001; 9643 onChanged(); 9644 return getJidFieldBuilder().getBuilder(); 9645 } 9646 /** 9647 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9648 */ 9649 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 9650 if (jidBuilder_ != null) { 9651 return jidBuilder_.getMessageOrBuilder(); 9652 } else { 9653 return jid_; 9654 } 9655 } 9656 /** 9657 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 9658 */ 9659 private com.google.protobuf.SingleFieldBuilder< 9660 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 9661 getJidFieldBuilder() { 9662 if (jidBuilder_ == null) { 9663 jidBuilder_ = new com.google.protobuf.SingleFieldBuilder< 9664 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 9665 jid_, 9666 getParentForChildren(), 9667 isClean()); 9668 jid_ = null; 9669 } 9670 return jidBuilder_; 9671 } 9672 9673 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateRequestProto) 9674 } 9675 9676 static { 9677 defaultInstance = new GetJournalStateRequestProto(true); 9678 defaultInstance.initFields(); 9679 } 9680 9681 // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateRequestProto) 9682 } 9683 9684 public interface GetJournalStateResponseProtoOrBuilder 9685 extends com.google.protobuf.MessageOrBuilder { 9686 9687 // required uint64 lastPromisedEpoch = 1; 9688 /** 9689 * <code>required uint64 lastPromisedEpoch = 1;</code> 9690 */ 9691 boolean hasLastPromisedEpoch(); 9692 /** 9693 * <code>required uint64 lastPromisedEpoch = 1;</code> 9694 */ 9695 long getLastPromisedEpoch(); 9696 9697 // required uint32 httpPort = 2; 9698 /** 9699 * <code>required uint32 httpPort = 2;</code> 9700 * 9701 * <pre> 9702 * Deprecated by fromURL 9703 * </pre> 9704 */ 9705 boolean hasHttpPort(); 9706 /** 9707 * <code>required uint32 httpPort = 2;</code> 9708 * 9709 * <pre> 9710 * Deprecated by fromURL 9711 * </pre> 9712 */ 9713 int getHttpPort(); 9714 9715 // optional string fromURL = 3; 9716 /** 9717 * <code>optional string fromURL = 3;</code> 9718 */ 9719 boolean hasFromURL(); 9720 /** 9721 * <code>optional string fromURL = 3;</code> 9722 */ 9723 java.lang.String getFromURL(); 9724 /** 9725 * <code>optional string fromURL = 3;</code> 9726 */ 9727 com.google.protobuf.ByteString 9728 getFromURLBytes(); 9729 } 9730 /** 9731 * Protobuf type {@code hadoop.hdfs.GetJournalStateResponseProto} 9732 */ 9733 public static final class GetJournalStateResponseProto extends 9734 com.google.protobuf.GeneratedMessage 9735 implements GetJournalStateResponseProtoOrBuilder { 9736 // Use GetJournalStateResponseProto.newBuilder() to construct. 9737 private GetJournalStateResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 9738 super(builder); 9739 this.unknownFields = builder.getUnknownFields(); 9740 } 9741 private GetJournalStateResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 9742 9743 private static final GetJournalStateResponseProto defaultInstance; 9744 public static GetJournalStateResponseProto getDefaultInstance() { 9745 return defaultInstance; 9746 } 9747 9748 public GetJournalStateResponseProto getDefaultInstanceForType() { 9749 return defaultInstance; 9750 } 9751 9752 private final com.google.protobuf.UnknownFieldSet unknownFields; 9753 @java.lang.Override 9754 public final com.google.protobuf.UnknownFieldSet 9755 getUnknownFields() { 9756 return this.unknownFields; 9757 } 9758 private GetJournalStateResponseProto( 9759 com.google.protobuf.CodedInputStream input, 9760 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9761 throws com.google.protobuf.InvalidProtocolBufferException { 9762 initFields(); 9763 int mutable_bitField0_ = 0; 9764 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 9765 com.google.protobuf.UnknownFieldSet.newBuilder(); 9766 try { 9767 boolean done = false; 9768 while (!done) { 9769 int tag = input.readTag(); 9770 switch (tag) { 9771 case 0: 9772 done = true; 9773 break; 9774 default: { 9775 if (!parseUnknownField(input, unknownFields, 9776 extensionRegistry, tag)) { 9777 done = true; 9778 } 9779 break; 9780 } 9781 case 8: { 9782 bitField0_ |= 0x00000001; 9783 lastPromisedEpoch_ = input.readUInt64(); 9784 break; 9785 } 9786 case 16: { 9787 bitField0_ |= 0x00000002; 9788 httpPort_ = input.readUInt32(); 9789 break; 9790 } 9791 case 26: { 9792 bitField0_ |= 0x00000004; 9793 fromURL_ = input.readBytes(); 9794 break; 9795 } 9796 } 9797 } 9798 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9799 throw e.setUnfinishedMessage(this); 9800 } catch (java.io.IOException e) { 9801 throw new com.google.protobuf.InvalidProtocolBufferException( 9802 e.getMessage()).setUnfinishedMessage(this); 9803 } finally { 9804 this.unknownFields = unknownFields.build(); 9805 makeExtensionsImmutable(); 9806 } 9807 } 9808 public static final com.google.protobuf.Descriptors.Descriptor 9809 getDescriptor() { 9810 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor; 9811 } 9812 9813 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9814 internalGetFieldAccessorTable() { 9815 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable 9816 .ensureFieldAccessorsInitialized( 9817 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class); 9818 } 9819 9820 public static com.google.protobuf.Parser<GetJournalStateResponseProto> PARSER = 9821 new com.google.protobuf.AbstractParser<GetJournalStateResponseProto>() { 9822 public GetJournalStateResponseProto parsePartialFrom( 9823 com.google.protobuf.CodedInputStream input, 9824 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9825 throws com.google.protobuf.InvalidProtocolBufferException { 9826 return new GetJournalStateResponseProto(input, extensionRegistry); 9827 } 9828 }; 9829 9830 @java.lang.Override 9831 public com.google.protobuf.Parser<GetJournalStateResponseProto> getParserForType() { 9832 return PARSER; 9833 } 9834 9835 private int bitField0_; 9836 // required uint64 lastPromisedEpoch = 1; 9837 public static final int LASTPROMISEDEPOCH_FIELD_NUMBER = 1; 9838 private long lastPromisedEpoch_; 9839 /** 9840 * <code>required uint64 lastPromisedEpoch = 1;</code> 9841 */ 9842 public boolean hasLastPromisedEpoch() { 9843 return ((bitField0_ & 0x00000001) == 0x00000001); 9844 } 9845 /** 9846 * <code>required uint64 lastPromisedEpoch = 1;</code> 9847 */ 9848 public long getLastPromisedEpoch() { 9849 return lastPromisedEpoch_; 9850 } 9851 9852 // required uint32 httpPort = 2; 9853 public static final int HTTPPORT_FIELD_NUMBER = 2; 9854 private int httpPort_; 9855 /** 9856 * <code>required uint32 httpPort = 2;</code> 9857 * 9858 * <pre> 9859 * Deprecated by fromURL 9860 * </pre> 9861 */ 9862 public boolean hasHttpPort() { 9863 return ((bitField0_ & 0x00000002) == 0x00000002); 9864 } 9865 /** 9866 * <code>required uint32 httpPort = 2;</code> 9867 * 9868 * <pre> 9869 * Deprecated by fromURL 9870 * </pre> 9871 */ 9872 public int getHttpPort() { 9873 return httpPort_; 9874 } 9875 9876 // optional string fromURL = 3; 9877 public static final int FROMURL_FIELD_NUMBER = 3; 9878 private java.lang.Object fromURL_; 9879 /** 9880 * <code>optional string fromURL = 3;</code> 9881 */ 9882 public boolean hasFromURL() { 9883 return ((bitField0_ & 0x00000004) == 0x00000004); 9884 } 9885 /** 9886 * <code>optional string fromURL = 3;</code> 9887 */ 9888 public java.lang.String getFromURL() { 9889 java.lang.Object ref = fromURL_; 9890 if (ref instanceof java.lang.String) { 9891 return (java.lang.String) ref; 9892 } else { 9893 com.google.protobuf.ByteString bs = 9894 (com.google.protobuf.ByteString) ref; 9895 java.lang.String s = bs.toStringUtf8(); 9896 if (bs.isValidUtf8()) { 9897 fromURL_ = s; 9898 } 9899 return s; 9900 } 9901 } 9902 /** 9903 * <code>optional string fromURL = 3;</code> 9904 */ 9905 public com.google.protobuf.ByteString 9906 getFromURLBytes() { 9907 java.lang.Object ref = fromURL_; 9908 if (ref instanceof java.lang.String) { 9909 com.google.protobuf.ByteString b = 9910 com.google.protobuf.ByteString.copyFromUtf8( 9911 (java.lang.String) ref); 9912 fromURL_ = b; 9913 return b; 9914 } else { 9915 return (com.google.protobuf.ByteString) ref; 9916 } 9917 } 9918 9919 private void initFields() { 9920 lastPromisedEpoch_ = 0L; 9921 httpPort_ = 0; 9922 fromURL_ = ""; 9923 } 9924 private byte memoizedIsInitialized = -1; 9925 public final boolean isInitialized() { 9926 byte isInitialized = memoizedIsInitialized; 9927 if (isInitialized != -1) return isInitialized == 1; 9928 9929 if (!hasLastPromisedEpoch()) { 9930 memoizedIsInitialized = 0; 9931 return false; 9932 } 9933 if (!hasHttpPort()) { 9934 memoizedIsInitialized = 0; 9935 return false; 9936 } 9937 memoizedIsInitialized = 1; 9938 return true; 9939 } 9940 9941 public void writeTo(com.google.protobuf.CodedOutputStream output) 9942 throws java.io.IOException { 9943 getSerializedSize(); 9944 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9945 output.writeUInt64(1, lastPromisedEpoch_); 9946 } 9947 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9948 output.writeUInt32(2, httpPort_); 9949 } 9950 if (((bitField0_ & 0x00000004) == 0x00000004)) { 9951 output.writeBytes(3, getFromURLBytes()); 9952 } 9953 getUnknownFields().writeTo(output); 9954 } 9955 9956 private int memoizedSerializedSize = -1; 9957 public int getSerializedSize() { 9958 int size = memoizedSerializedSize; 9959 if (size != -1) return size; 9960 9961 size = 0; 9962 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9963 size += com.google.protobuf.CodedOutputStream 9964 .computeUInt64Size(1, lastPromisedEpoch_); 9965 } 9966 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9967 size += com.google.protobuf.CodedOutputStream 9968 .computeUInt32Size(2, httpPort_); 9969 } 9970 if (((bitField0_ & 0x00000004) == 0x00000004)) { 9971 size += com.google.protobuf.CodedOutputStream 9972 .computeBytesSize(3, getFromURLBytes()); 9973 } 9974 size += getUnknownFields().getSerializedSize(); 9975 memoizedSerializedSize = size; 9976 return size; 9977 } 9978 9979 private static final long serialVersionUID = 0L; 9980 @java.lang.Override 9981 protected java.lang.Object writeReplace() 9982 throws java.io.ObjectStreamException { 9983 return super.writeReplace(); 9984 } 9985 9986 @java.lang.Override 9987 public boolean equals(final java.lang.Object obj) { 9988 if (obj == this) { 9989 return true; 9990 } 9991 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)) { 9992 return super.equals(obj); 9993 } 9994 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) obj; 9995 9996 boolean result = true; 9997 result = result && (hasLastPromisedEpoch() == other.hasLastPromisedEpoch()); 9998 if (hasLastPromisedEpoch()) { 9999 result = result && (getLastPromisedEpoch() 10000 == other.getLastPromisedEpoch()); 10001 } 10002 result = result && (hasHttpPort() == other.hasHttpPort()); 10003 if (hasHttpPort()) { 10004 result = result && (getHttpPort() 10005 == other.getHttpPort()); 10006 } 10007 result = result && (hasFromURL() == other.hasFromURL()); 10008 if (hasFromURL()) { 10009 result = result && getFromURL() 10010 .equals(other.getFromURL()); 10011 } 10012 result = result && 10013 getUnknownFields().equals(other.getUnknownFields()); 10014 return result; 10015 } 10016 10017 private int memoizedHashCode = 0; 10018 @java.lang.Override 10019 public int hashCode() { 10020 if (memoizedHashCode != 0) { 10021 return memoizedHashCode; 10022 } 10023 int hash = 41; 10024 hash = (19 * hash) + getDescriptorForType().hashCode(); 10025 if (hasLastPromisedEpoch()) { 10026 hash = (37 * hash) + LASTPROMISEDEPOCH_FIELD_NUMBER; 10027 hash = (53 * hash) + hashLong(getLastPromisedEpoch()); 10028 } 10029 if (hasHttpPort()) { 10030 hash = (37 * hash) + HTTPPORT_FIELD_NUMBER; 10031 hash = (53 * hash) + getHttpPort(); 10032 } 10033 if (hasFromURL()) { 10034 hash = (37 * hash) + FROMURL_FIELD_NUMBER; 10035 hash = (53 * hash) + getFromURL().hashCode(); 10036 } 10037 hash = (29 * hash) + getUnknownFields().hashCode(); 10038 memoizedHashCode = hash; 10039 return hash; 10040 } 10041 10042 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10043 com.google.protobuf.ByteString data) 10044 throws com.google.protobuf.InvalidProtocolBufferException { 10045 return PARSER.parseFrom(data); 10046 } 10047 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10048 com.google.protobuf.ByteString data, 10049 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10050 throws com.google.protobuf.InvalidProtocolBufferException { 10051 return PARSER.parseFrom(data, extensionRegistry); 10052 } 10053 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(byte[] data) 10054 throws com.google.protobuf.InvalidProtocolBufferException { 10055 return PARSER.parseFrom(data); 10056 } 10057 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10058 byte[] data, 10059 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10060 throws com.google.protobuf.InvalidProtocolBufferException { 10061 return PARSER.parseFrom(data, extensionRegistry); 10062 } 10063 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(java.io.InputStream input) 10064 throws java.io.IOException { 10065 return PARSER.parseFrom(input); 10066 } 10067 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10068 java.io.InputStream input, 10069 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10070 throws java.io.IOException { 10071 return PARSER.parseFrom(input, extensionRegistry); 10072 } 10073 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(java.io.InputStream input) 10074 throws java.io.IOException { 10075 return PARSER.parseDelimitedFrom(input); 10076 } 10077 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom( 10078 java.io.InputStream input, 10079 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10080 throws java.io.IOException { 10081 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10082 } 10083 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10084 com.google.protobuf.CodedInputStream input) 10085 throws java.io.IOException { 10086 return PARSER.parseFrom(input); 10087 } 10088 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom( 10089 com.google.protobuf.CodedInputStream input, 10090 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10091 throws java.io.IOException { 10092 return PARSER.parseFrom(input, extensionRegistry); 10093 } 10094 10095 public static Builder newBuilder() { return Builder.create(); } 10096 public Builder newBuilderForType() { return newBuilder(); } 10097 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto prototype) { 10098 return newBuilder().mergeFrom(prototype); 10099 } 10100 public Builder toBuilder() { return newBuilder(this); } 10101 10102 @java.lang.Override 10103 protected Builder newBuilderForType( 10104 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10105 Builder builder = new Builder(parent); 10106 return builder; 10107 } 10108 /** 10109 * Protobuf type {@code hadoop.hdfs.GetJournalStateResponseProto} 10110 */ 10111 public static final class Builder extends 10112 com.google.protobuf.GeneratedMessage.Builder<Builder> 10113 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProtoOrBuilder { 10114 public static final com.google.protobuf.Descriptors.Descriptor 10115 getDescriptor() { 10116 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor; 10117 } 10118 10119 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10120 internalGetFieldAccessorTable() { 10121 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable 10122 .ensureFieldAccessorsInitialized( 10123 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class); 10124 } 10125 10126 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.newBuilder() 10127 private Builder() { 10128 maybeForceBuilderInitialization(); 10129 } 10130 10131 private Builder( 10132 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10133 super(parent); 10134 maybeForceBuilderInitialization(); 10135 } 10136 private void maybeForceBuilderInitialization() { 10137 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10138 } 10139 } 10140 private static Builder create() { 10141 return new Builder(); 10142 } 10143 10144 public Builder clear() { 10145 super.clear(); 10146 lastPromisedEpoch_ = 0L; 10147 bitField0_ = (bitField0_ & ~0x00000001); 10148 httpPort_ = 0; 10149 bitField0_ = (bitField0_ & ~0x00000002); 10150 fromURL_ = ""; 10151 bitField0_ = (bitField0_ & ~0x00000004); 10152 return this; 10153 } 10154 10155 public Builder clone() { 10156 return create().mergeFrom(buildPartial()); 10157 } 10158 10159 public com.google.protobuf.Descriptors.Descriptor 10160 getDescriptorForType() { 10161 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor; 10162 } 10163 10164 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getDefaultInstanceForType() { 10165 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(); 10166 } 10167 10168 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto build() { 10169 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = buildPartial(); 10170 if (!result.isInitialized()) { 10171 throw newUninitializedMessageException(result); 10172 } 10173 return result; 10174 } 10175 10176 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto buildPartial() { 10177 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto(this); 10178 int from_bitField0_ = bitField0_; 10179 int to_bitField0_ = 0; 10180 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 10181 to_bitField0_ |= 0x00000001; 10182 } 10183 result.lastPromisedEpoch_ = lastPromisedEpoch_; 10184 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 10185 to_bitField0_ |= 0x00000002; 10186 } 10187 result.httpPort_ = httpPort_; 10188 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 10189 to_bitField0_ |= 0x00000004; 10190 } 10191 result.fromURL_ = fromURL_; 10192 result.bitField0_ = to_bitField0_; 10193 onBuilt(); 10194 return result; 10195 } 10196 10197 public Builder mergeFrom(com.google.protobuf.Message other) { 10198 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) { 10199 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)other); 10200 } else { 10201 super.mergeFrom(other); 10202 return this; 10203 } 10204 } 10205 10206 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other) { 10207 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()) return this; 10208 if (other.hasLastPromisedEpoch()) { 10209 setLastPromisedEpoch(other.getLastPromisedEpoch()); 10210 } 10211 if (other.hasHttpPort()) { 10212 setHttpPort(other.getHttpPort()); 10213 } 10214 if (other.hasFromURL()) { 10215 bitField0_ |= 0x00000004; 10216 fromURL_ = other.fromURL_; 10217 onChanged(); 10218 } 10219 this.mergeUnknownFields(other.getUnknownFields()); 10220 return this; 10221 } 10222 10223 public final boolean isInitialized() { 10224 if (!hasLastPromisedEpoch()) { 10225 10226 return false; 10227 } 10228 if (!hasHttpPort()) { 10229 10230 return false; 10231 } 10232 return true; 10233 } 10234 10235 public Builder mergeFrom( 10236 com.google.protobuf.CodedInputStream input, 10237 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10238 throws java.io.IOException { 10239 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parsedMessage = null; 10240 try { 10241 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 10242 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10243 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) e.getUnfinishedMessage(); 10244 throw e; 10245 } finally { 10246 if (parsedMessage != null) { 10247 mergeFrom(parsedMessage); 10248 } 10249 } 10250 return this; 10251 } 10252 private int bitField0_; 10253 10254 // required uint64 lastPromisedEpoch = 1; 10255 private long lastPromisedEpoch_ ; 10256 /** 10257 * <code>required uint64 lastPromisedEpoch = 1;</code> 10258 */ 10259 public boolean hasLastPromisedEpoch() { 10260 return ((bitField0_ & 0x00000001) == 0x00000001); 10261 } 10262 /** 10263 * <code>required uint64 lastPromisedEpoch = 1;</code> 10264 */ 10265 public long getLastPromisedEpoch() { 10266 return lastPromisedEpoch_; 10267 } 10268 /** 10269 * <code>required uint64 lastPromisedEpoch = 1;</code> 10270 */ 10271 public Builder setLastPromisedEpoch(long value) { 10272 bitField0_ |= 0x00000001; 10273 lastPromisedEpoch_ = value; 10274 onChanged(); 10275 return this; 10276 } 10277 /** 10278 * <code>required uint64 lastPromisedEpoch = 1;</code> 10279 */ 10280 public Builder clearLastPromisedEpoch() { 10281 bitField0_ = (bitField0_ & ~0x00000001); 10282 lastPromisedEpoch_ = 0L; 10283 onChanged(); 10284 return this; 10285 } 10286 10287 // required uint32 httpPort = 2; 10288 private int httpPort_ ; 10289 /** 10290 * <code>required uint32 httpPort = 2;</code> 10291 * 10292 * <pre> 10293 * Deprecated by fromURL 10294 * </pre> 10295 */ 10296 public boolean hasHttpPort() { 10297 return ((bitField0_ & 0x00000002) == 0x00000002); 10298 } 10299 /** 10300 * <code>required uint32 httpPort = 2;</code> 10301 * 10302 * <pre> 10303 * Deprecated by fromURL 10304 * </pre> 10305 */ 10306 public int getHttpPort() { 10307 return httpPort_; 10308 } 10309 /** 10310 * <code>required uint32 httpPort = 2;</code> 10311 * 10312 * <pre> 10313 * Deprecated by fromURL 10314 * </pre> 10315 */ 10316 public Builder setHttpPort(int value) { 10317 bitField0_ |= 0x00000002; 10318 httpPort_ = value; 10319 onChanged(); 10320 return this; 10321 } 10322 /** 10323 * <code>required uint32 httpPort = 2;</code> 10324 * 10325 * <pre> 10326 * Deprecated by fromURL 10327 * </pre> 10328 */ 10329 public Builder clearHttpPort() { 10330 bitField0_ = (bitField0_ & ~0x00000002); 10331 httpPort_ = 0; 10332 onChanged(); 10333 return this; 10334 } 10335 10336 // optional string fromURL = 3; 10337 private java.lang.Object fromURL_ = ""; 10338 /** 10339 * <code>optional string fromURL = 3;</code> 10340 */ 10341 public boolean hasFromURL() { 10342 return ((bitField0_ & 0x00000004) == 0x00000004); 10343 } 10344 /** 10345 * <code>optional string fromURL = 3;</code> 10346 */ 10347 public java.lang.String getFromURL() { 10348 java.lang.Object ref = fromURL_; 10349 if (!(ref instanceof java.lang.String)) { 10350 java.lang.String s = ((com.google.protobuf.ByteString) ref) 10351 .toStringUtf8(); 10352 fromURL_ = s; 10353 return s; 10354 } else { 10355 return (java.lang.String) ref; 10356 } 10357 } 10358 /** 10359 * <code>optional string fromURL = 3;</code> 10360 */ 10361 public com.google.protobuf.ByteString 10362 getFromURLBytes() { 10363 java.lang.Object ref = fromURL_; 10364 if (ref instanceof String) { 10365 com.google.protobuf.ByteString b = 10366 com.google.protobuf.ByteString.copyFromUtf8( 10367 (java.lang.String) ref); 10368 fromURL_ = b; 10369 return b; 10370 } else { 10371 return (com.google.protobuf.ByteString) ref; 10372 } 10373 } 10374 /** 10375 * <code>optional string fromURL = 3;</code> 10376 */ 10377 public Builder setFromURL( 10378 java.lang.String value) { 10379 if (value == null) { 10380 throw new NullPointerException(); 10381 } 10382 bitField0_ |= 0x00000004; 10383 fromURL_ = value; 10384 onChanged(); 10385 return this; 10386 } 10387 /** 10388 * <code>optional string fromURL = 3;</code> 10389 */ 10390 public Builder clearFromURL() { 10391 bitField0_ = (bitField0_ & ~0x00000004); 10392 fromURL_ = getDefaultInstance().getFromURL(); 10393 onChanged(); 10394 return this; 10395 } 10396 /** 10397 * <code>optional string fromURL = 3;</code> 10398 */ 10399 public Builder setFromURLBytes( 10400 com.google.protobuf.ByteString value) { 10401 if (value == null) { 10402 throw new NullPointerException(); 10403 } 10404 bitField0_ |= 0x00000004; 10405 fromURL_ = value; 10406 onChanged(); 10407 return this; 10408 } 10409 10410 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateResponseProto) 10411 } 10412 10413 static { 10414 defaultInstance = new GetJournalStateResponseProto(true); 10415 defaultInstance.initFields(); 10416 } 10417 10418 // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateResponseProto) 10419 } 10420 10421 public interface FormatRequestProtoOrBuilder 10422 extends com.google.protobuf.MessageOrBuilder { 10423 10424 // required .hadoop.hdfs.JournalIdProto jid = 1; 10425 /** 10426 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10427 */ 10428 boolean hasJid(); 10429 /** 10430 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10431 */ 10432 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid(); 10433 /** 10434 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10435 */ 10436 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder(); 10437 10438 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 10439 /** 10440 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10441 */ 10442 boolean hasNsInfo(); 10443 /** 10444 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10445 */ 10446 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo(); 10447 /** 10448 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10449 */ 10450 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder(); 10451 } 10452 /** 10453 * Protobuf type {@code hadoop.hdfs.FormatRequestProto} 10454 * 10455 * <pre> 10456 ** 10457 * format() 10458 * </pre> 10459 */ 10460 public static final class FormatRequestProto extends 10461 com.google.protobuf.GeneratedMessage 10462 implements FormatRequestProtoOrBuilder { 10463 // Use FormatRequestProto.newBuilder() to construct. 10464 private FormatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 10465 super(builder); 10466 this.unknownFields = builder.getUnknownFields(); 10467 } 10468 private FormatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 10469 10470 private static final FormatRequestProto defaultInstance; 10471 public static FormatRequestProto getDefaultInstance() { 10472 return defaultInstance; 10473 } 10474 10475 public FormatRequestProto getDefaultInstanceForType() { 10476 return defaultInstance; 10477 } 10478 10479 private final com.google.protobuf.UnknownFieldSet unknownFields; 10480 @java.lang.Override 10481 public final com.google.protobuf.UnknownFieldSet 10482 getUnknownFields() { 10483 return this.unknownFields; 10484 } 10485 private FormatRequestProto( 10486 com.google.protobuf.CodedInputStream input, 10487 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10488 throws com.google.protobuf.InvalidProtocolBufferException { 10489 initFields(); 10490 int mutable_bitField0_ = 0; 10491 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 10492 com.google.protobuf.UnknownFieldSet.newBuilder(); 10493 try { 10494 boolean done = false; 10495 while (!done) { 10496 int tag = input.readTag(); 10497 switch (tag) { 10498 case 0: 10499 done = true; 10500 break; 10501 default: { 10502 if (!parseUnknownField(input, unknownFields, 10503 extensionRegistry, tag)) { 10504 done = true; 10505 } 10506 break; 10507 } 10508 case 10: { 10509 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 10510 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10511 subBuilder = jid_.toBuilder(); 10512 } 10513 jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 10514 if (subBuilder != null) { 10515 subBuilder.mergeFrom(jid_); 10516 jid_ = subBuilder.buildPartial(); 10517 } 10518 bitField0_ |= 0x00000001; 10519 break; 10520 } 10521 case 18: { 10522 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = null; 10523 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10524 subBuilder = nsInfo_.toBuilder(); 10525 } 10526 nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.PARSER, extensionRegistry); 10527 if (subBuilder != null) { 10528 subBuilder.mergeFrom(nsInfo_); 10529 nsInfo_ = subBuilder.buildPartial(); 10530 } 10531 bitField0_ |= 0x00000002; 10532 break; 10533 } 10534 } 10535 } 10536 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10537 throw e.setUnfinishedMessage(this); 10538 } catch (java.io.IOException e) { 10539 throw new com.google.protobuf.InvalidProtocolBufferException( 10540 e.getMessage()).setUnfinishedMessage(this); 10541 } finally { 10542 this.unknownFields = unknownFields.build(); 10543 makeExtensionsImmutable(); 10544 } 10545 } 10546 public static final com.google.protobuf.Descriptors.Descriptor 10547 getDescriptor() { 10548 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor; 10549 } 10550 10551 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10552 internalGetFieldAccessorTable() { 10553 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable 10554 .ensureFieldAccessorsInitialized( 10555 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class); 10556 } 10557 10558 public static com.google.protobuf.Parser<FormatRequestProto> PARSER = 10559 new com.google.protobuf.AbstractParser<FormatRequestProto>() { 10560 public FormatRequestProto parsePartialFrom( 10561 com.google.protobuf.CodedInputStream input, 10562 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10563 throws com.google.protobuf.InvalidProtocolBufferException { 10564 return new FormatRequestProto(input, extensionRegistry); 10565 } 10566 }; 10567 10568 @java.lang.Override 10569 public com.google.protobuf.Parser<FormatRequestProto> getParserForType() { 10570 return PARSER; 10571 } 10572 10573 private int bitField0_; 10574 // required .hadoop.hdfs.JournalIdProto jid = 1; 10575 public static final int JID_FIELD_NUMBER = 1; 10576 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_; 10577 /** 10578 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10579 */ 10580 public boolean hasJid() { 10581 return ((bitField0_ & 0x00000001) == 0x00000001); 10582 } 10583 /** 10584 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10585 */ 10586 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 10587 return jid_; 10588 } 10589 /** 10590 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10591 */ 10592 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 10593 return jid_; 10594 } 10595 10596 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 10597 public static final int NSINFO_FIELD_NUMBER = 2; 10598 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_; 10599 /** 10600 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10601 */ 10602 public boolean hasNsInfo() { 10603 return ((bitField0_ & 0x00000002) == 0x00000002); 10604 } 10605 /** 10606 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10607 */ 10608 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() { 10609 return nsInfo_; 10610 } 10611 /** 10612 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 10613 */ 10614 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() { 10615 return nsInfo_; 10616 } 10617 10618 private void initFields() { 10619 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 10620 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 10621 } 10622 private byte memoizedIsInitialized = -1; 10623 public final boolean isInitialized() { 10624 byte isInitialized = memoizedIsInitialized; 10625 if (isInitialized != -1) return isInitialized == 1; 10626 10627 if (!hasJid()) { 10628 memoizedIsInitialized = 0; 10629 return false; 10630 } 10631 if (!hasNsInfo()) { 10632 memoizedIsInitialized = 0; 10633 return false; 10634 } 10635 if (!getJid().isInitialized()) { 10636 memoizedIsInitialized = 0; 10637 return false; 10638 } 10639 if (!getNsInfo().isInitialized()) { 10640 memoizedIsInitialized = 0; 10641 return false; 10642 } 10643 memoizedIsInitialized = 1; 10644 return true; 10645 } 10646 10647 public void writeTo(com.google.protobuf.CodedOutputStream output) 10648 throws java.io.IOException { 10649 getSerializedSize(); 10650 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10651 output.writeMessage(1, jid_); 10652 } 10653 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10654 output.writeMessage(2, nsInfo_); 10655 } 10656 getUnknownFields().writeTo(output); 10657 } 10658 10659 private int memoizedSerializedSize = -1; 10660 public int getSerializedSize() { 10661 int size = memoizedSerializedSize; 10662 if (size != -1) return size; 10663 10664 size = 0; 10665 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10666 size += com.google.protobuf.CodedOutputStream 10667 .computeMessageSize(1, jid_); 10668 } 10669 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10670 size += com.google.protobuf.CodedOutputStream 10671 .computeMessageSize(2, nsInfo_); 10672 } 10673 size += getUnknownFields().getSerializedSize(); 10674 memoizedSerializedSize = size; 10675 return size; 10676 } 10677 10678 private static final long serialVersionUID = 0L; 10679 @java.lang.Override 10680 protected java.lang.Object writeReplace() 10681 throws java.io.ObjectStreamException { 10682 return super.writeReplace(); 10683 } 10684 10685 @java.lang.Override 10686 public boolean equals(final java.lang.Object obj) { 10687 if (obj == this) { 10688 return true; 10689 } 10690 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)) { 10691 return super.equals(obj); 10692 } 10693 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) obj; 10694 10695 boolean result = true; 10696 result = result && (hasJid() == other.hasJid()); 10697 if (hasJid()) { 10698 result = result && getJid() 10699 .equals(other.getJid()); 10700 } 10701 result = result && (hasNsInfo() == other.hasNsInfo()); 10702 if (hasNsInfo()) { 10703 result = result && getNsInfo() 10704 .equals(other.getNsInfo()); 10705 } 10706 result = result && 10707 getUnknownFields().equals(other.getUnknownFields()); 10708 return result; 10709 } 10710 10711 private int memoizedHashCode = 0; 10712 @java.lang.Override 10713 public int hashCode() { 10714 if (memoizedHashCode != 0) { 10715 return memoizedHashCode; 10716 } 10717 int hash = 41; 10718 hash = (19 * hash) + getDescriptorForType().hashCode(); 10719 if (hasJid()) { 10720 hash = (37 * hash) + JID_FIELD_NUMBER; 10721 hash = (53 * hash) + getJid().hashCode(); 10722 } 10723 if (hasNsInfo()) { 10724 hash = (37 * hash) + NSINFO_FIELD_NUMBER; 10725 hash = (53 * hash) + getNsInfo().hashCode(); 10726 } 10727 hash = (29 * hash) + getUnknownFields().hashCode(); 10728 memoizedHashCode = hash; 10729 return hash; 10730 } 10731 10732 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10733 com.google.protobuf.ByteString data) 10734 throws com.google.protobuf.InvalidProtocolBufferException { 10735 return PARSER.parseFrom(data); 10736 } 10737 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10738 com.google.protobuf.ByteString data, 10739 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10740 throws com.google.protobuf.InvalidProtocolBufferException { 10741 return PARSER.parseFrom(data, extensionRegistry); 10742 } 10743 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(byte[] data) 10744 throws com.google.protobuf.InvalidProtocolBufferException { 10745 return PARSER.parseFrom(data); 10746 } 10747 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10748 byte[] data, 10749 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10750 throws com.google.protobuf.InvalidProtocolBufferException { 10751 return PARSER.parseFrom(data, extensionRegistry); 10752 } 10753 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(java.io.InputStream input) 10754 throws java.io.IOException { 10755 return PARSER.parseFrom(input); 10756 } 10757 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10758 java.io.InputStream input, 10759 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10760 throws java.io.IOException { 10761 return PARSER.parseFrom(input, extensionRegistry); 10762 } 10763 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(java.io.InputStream input) 10764 throws java.io.IOException { 10765 return PARSER.parseDelimitedFrom(input); 10766 } 10767 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom( 10768 java.io.InputStream input, 10769 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10770 throws java.io.IOException { 10771 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10772 } 10773 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10774 com.google.protobuf.CodedInputStream input) 10775 throws java.io.IOException { 10776 return PARSER.parseFrom(input); 10777 } 10778 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom( 10779 com.google.protobuf.CodedInputStream input, 10780 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10781 throws java.io.IOException { 10782 return PARSER.parseFrom(input, extensionRegistry); 10783 } 10784 10785 public static Builder newBuilder() { return Builder.create(); } 10786 public Builder newBuilderForType() { return newBuilder(); } 10787 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto prototype) { 10788 return newBuilder().mergeFrom(prototype); 10789 } 10790 public Builder toBuilder() { return newBuilder(this); } 10791 10792 @java.lang.Override 10793 protected Builder newBuilderForType( 10794 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10795 Builder builder = new Builder(parent); 10796 return builder; 10797 } 10798 /** 10799 * Protobuf type {@code hadoop.hdfs.FormatRequestProto} 10800 * 10801 * <pre> 10802 ** 10803 * format() 10804 * </pre> 10805 */ 10806 public static final class Builder extends 10807 com.google.protobuf.GeneratedMessage.Builder<Builder> 10808 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProtoOrBuilder { 10809 public static final com.google.protobuf.Descriptors.Descriptor 10810 getDescriptor() { 10811 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor; 10812 } 10813 10814 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10815 internalGetFieldAccessorTable() { 10816 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable 10817 .ensureFieldAccessorsInitialized( 10818 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class); 10819 } 10820 10821 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.newBuilder() 10822 private Builder() { 10823 maybeForceBuilderInitialization(); 10824 } 10825 10826 private Builder( 10827 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10828 super(parent); 10829 maybeForceBuilderInitialization(); 10830 } 10831 private void maybeForceBuilderInitialization() { 10832 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10833 getJidFieldBuilder(); 10834 getNsInfoFieldBuilder(); 10835 } 10836 } 10837 private static Builder create() { 10838 return new Builder(); 10839 } 10840 10841 public Builder clear() { 10842 super.clear(); 10843 if (jidBuilder_ == null) { 10844 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 10845 } else { 10846 jidBuilder_.clear(); 10847 } 10848 bitField0_ = (bitField0_ & ~0x00000001); 10849 if (nsInfoBuilder_ == null) { 10850 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 10851 } else { 10852 nsInfoBuilder_.clear(); 10853 } 10854 bitField0_ = (bitField0_ & ~0x00000002); 10855 return this; 10856 } 10857 10858 public Builder clone() { 10859 return create().mergeFrom(buildPartial()); 10860 } 10861 10862 public com.google.protobuf.Descriptors.Descriptor 10863 getDescriptorForType() { 10864 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor; 10865 } 10866 10867 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto getDefaultInstanceForType() { 10868 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance(); 10869 } 10870 10871 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto build() { 10872 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = buildPartial(); 10873 if (!result.isInitialized()) { 10874 throw newUninitializedMessageException(result); 10875 } 10876 return result; 10877 } 10878 10879 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto buildPartial() { 10880 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto(this); 10881 int from_bitField0_ = bitField0_; 10882 int to_bitField0_ = 0; 10883 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 10884 to_bitField0_ |= 0x00000001; 10885 } 10886 if (jidBuilder_ == null) { 10887 result.jid_ = jid_; 10888 } else { 10889 result.jid_ = jidBuilder_.build(); 10890 } 10891 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 10892 to_bitField0_ |= 0x00000002; 10893 } 10894 if (nsInfoBuilder_ == null) { 10895 result.nsInfo_ = nsInfo_; 10896 } else { 10897 result.nsInfo_ = nsInfoBuilder_.build(); 10898 } 10899 result.bitField0_ = to_bitField0_; 10900 onBuilt(); 10901 return result; 10902 } 10903 10904 public Builder mergeFrom(com.google.protobuf.Message other) { 10905 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) { 10906 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)other); 10907 } else { 10908 super.mergeFrom(other); 10909 return this; 10910 } 10911 } 10912 10913 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other) { 10914 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance()) return this; 10915 if (other.hasJid()) { 10916 mergeJid(other.getJid()); 10917 } 10918 if (other.hasNsInfo()) { 10919 mergeNsInfo(other.getNsInfo()); 10920 } 10921 this.mergeUnknownFields(other.getUnknownFields()); 10922 return this; 10923 } 10924 10925 public final boolean isInitialized() { 10926 if (!hasJid()) { 10927 10928 return false; 10929 } 10930 if (!hasNsInfo()) { 10931 10932 return false; 10933 } 10934 if (!getJid().isInitialized()) { 10935 10936 return false; 10937 } 10938 if (!getNsInfo().isInitialized()) { 10939 10940 return false; 10941 } 10942 return true; 10943 } 10944 10945 public Builder mergeFrom( 10946 com.google.protobuf.CodedInputStream input, 10947 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10948 throws java.io.IOException { 10949 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parsedMessage = null; 10950 try { 10951 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 10952 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10953 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) e.getUnfinishedMessage(); 10954 throw e; 10955 } finally { 10956 if (parsedMessage != null) { 10957 mergeFrom(parsedMessage); 10958 } 10959 } 10960 return this; 10961 } 10962 private int bitField0_; 10963 10964 // required .hadoop.hdfs.JournalIdProto jid = 1; 10965 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 10966 private com.google.protobuf.SingleFieldBuilder< 10967 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_; 10968 /** 10969 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10970 */ 10971 public boolean hasJid() { 10972 return ((bitField0_ & 0x00000001) == 0x00000001); 10973 } 10974 /** 10975 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10976 */ 10977 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 10978 if (jidBuilder_ == null) { 10979 return jid_; 10980 } else { 10981 return jidBuilder_.getMessage(); 10982 } 10983 } 10984 /** 10985 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 10986 */ 10987 public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 10988 if (jidBuilder_ == null) { 10989 if (value == null) { 10990 throw new NullPointerException(); 10991 } 10992 jid_ = value; 10993 onChanged(); 10994 } else { 10995 jidBuilder_.setMessage(value); 10996 } 10997 bitField0_ |= 0x00000001; 10998 return this; 10999 } 11000 /** 11001 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11002 */ 11003 public Builder setJid( 11004 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 11005 if (jidBuilder_ == null) { 11006 jid_ = builderForValue.build(); 11007 onChanged(); 11008 } else { 11009 jidBuilder_.setMessage(builderForValue.build()); 11010 } 11011 bitField0_ |= 0x00000001; 11012 return this; 11013 } 11014 /** 11015 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11016 */ 11017 public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 11018 if (jidBuilder_ == null) { 11019 if (((bitField0_ & 0x00000001) == 0x00000001) && 11020 jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 11021 jid_ = 11022 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial(); 11023 } else { 11024 jid_ = value; 11025 } 11026 onChanged(); 11027 } else { 11028 jidBuilder_.mergeFrom(value); 11029 } 11030 bitField0_ |= 0x00000001; 11031 return this; 11032 } 11033 /** 11034 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11035 */ 11036 public Builder clearJid() { 11037 if (jidBuilder_ == null) { 11038 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 11039 onChanged(); 11040 } else { 11041 jidBuilder_.clear(); 11042 } 11043 bitField0_ = (bitField0_ & ~0x00000001); 11044 return this; 11045 } 11046 /** 11047 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11048 */ 11049 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() { 11050 bitField0_ |= 0x00000001; 11051 onChanged(); 11052 return getJidFieldBuilder().getBuilder(); 11053 } 11054 /** 11055 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11056 */ 11057 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 11058 if (jidBuilder_ != null) { 11059 return jidBuilder_.getMessageOrBuilder(); 11060 } else { 11061 return jid_; 11062 } 11063 } 11064 /** 11065 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11066 */ 11067 private com.google.protobuf.SingleFieldBuilder< 11068 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 11069 getJidFieldBuilder() { 11070 if (jidBuilder_ == null) { 11071 jidBuilder_ = new com.google.protobuf.SingleFieldBuilder< 11072 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 11073 jid_, 11074 getParentForChildren(), 11075 isClean()); 11076 jid_ = null; 11077 } 11078 return jidBuilder_; 11079 } 11080 11081 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 11082 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 11083 private com.google.protobuf.SingleFieldBuilder< 11084 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_; 11085 /** 11086 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11087 */ 11088 public boolean hasNsInfo() { 11089 return ((bitField0_ & 0x00000002) == 0x00000002); 11090 } 11091 /** 11092 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11093 */ 11094 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() { 11095 if (nsInfoBuilder_ == null) { 11096 return nsInfo_; 11097 } else { 11098 return nsInfoBuilder_.getMessage(); 11099 } 11100 } 11101 /** 11102 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11103 */ 11104 public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) { 11105 if (nsInfoBuilder_ == null) { 11106 if (value == null) { 11107 throw new NullPointerException(); 11108 } 11109 nsInfo_ = value; 11110 onChanged(); 11111 } else { 11112 nsInfoBuilder_.setMessage(value); 11113 } 11114 bitField0_ |= 0x00000002; 11115 return this; 11116 } 11117 /** 11118 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11119 */ 11120 public Builder setNsInfo( 11121 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) { 11122 if (nsInfoBuilder_ == null) { 11123 nsInfo_ = builderForValue.build(); 11124 onChanged(); 11125 } else { 11126 nsInfoBuilder_.setMessage(builderForValue.build()); 11127 } 11128 bitField0_ |= 0x00000002; 11129 return this; 11130 } 11131 /** 11132 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11133 */ 11134 public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) { 11135 if (nsInfoBuilder_ == null) { 11136 if (((bitField0_ & 0x00000002) == 0x00000002) && 11137 nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) { 11138 nsInfo_ = 11139 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial(); 11140 } else { 11141 nsInfo_ = value; 11142 } 11143 onChanged(); 11144 } else { 11145 nsInfoBuilder_.mergeFrom(value); 11146 } 11147 bitField0_ |= 0x00000002; 11148 return this; 11149 } 11150 /** 11151 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11152 */ 11153 public Builder clearNsInfo() { 11154 if (nsInfoBuilder_ == null) { 11155 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 11156 onChanged(); 11157 } else { 11158 nsInfoBuilder_.clear(); 11159 } 11160 bitField0_ = (bitField0_ & ~0x00000002); 11161 return this; 11162 } 11163 /** 11164 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11165 */ 11166 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() { 11167 bitField0_ |= 0x00000002; 11168 onChanged(); 11169 return getNsInfoFieldBuilder().getBuilder(); 11170 } 11171 /** 11172 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11173 */ 11174 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() { 11175 if (nsInfoBuilder_ != null) { 11176 return nsInfoBuilder_.getMessageOrBuilder(); 11177 } else { 11178 return nsInfo_; 11179 } 11180 } 11181 /** 11182 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11183 */ 11184 private com.google.protobuf.SingleFieldBuilder< 11185 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 11186 getNsInfoFieldBuilder() { 11187 if (nsInfoBuilder_ == null) { 11188 nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 11189 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>( 11190 nsInfo_, 11191 getParentForChildren(), 11192 isClean()); 11193 nsInfo_ = null; 11194 } 11195 return nsInfoBuilder_; 11196 } 11197 11198 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatRequestProto) 11199 } 11200 11201 static { 11202 defaultInstance = new FormatRequestProto(true); 11203 defaultInstance.initFields(); 11204 } 11205 11206 // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatRequestProto) 11207 } 11208 11209 public interface FormatResponseProtoOrBuilder 11210 extends com.google.protobuf.MessageOrBuilder { 11211 } 11212 /** 11213 * Protobuf type {@code hadoop.hdfs.FormatResponseProto} 11214 */ 11215 public static final class FormatResponseProto extends 11216 com.google.protobuf.GeneratedMessage 11217 implements FormatResponseProtoOrBuilder { 11218 // Use FormatResponseProto.newBuilder() to construct. 11219 private FormatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 11220 super(builder); 11221 this.unknownFields = builder.getUnknownFields(); 11222 } 11223 private FormatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 11224 11225 private static final FormatResponseProto defaultInstance; 11226 public static FormatResponseProto getDefaultInstance() { 11227 return defaultInstance; 11228 } 11229 11230 public FormatResponseProto getDefaultInstanceForType() { 11231 return defaultInstance; 11232 } 11233 11234 private final com.google.protobuf.UnknownFieldSet unknownFields; 11235 @java.lang.Override 11236 public final com.google.protobuf.UnknownFieldSet 11237 getUnknownFields() { 11238 return this.unknownFields; 11239 } 11240 private FormatResponseProto( 11241 com.google.protobuf.CodedInputStream input, 11242 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11243 throws com.google.protobuf.InvalidProtocolBufferException { 11244 initFields(); 11245 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 11246 com.google.protobuf.UnknownFieldSet.newBuilder(); 11247 try { 11248 boolean done = false; 11249 while (!done) { 11250 int tag = input.readTag(); 11251 switch (tag) { 11252 case 0: 11253 done = true; 11254 break; 11255 default: { 11256 if (!parseUnknownField(input, unknownFields, 11257 extensionRegistry, tag)) { 11258 done = true; 11259 } 11260 break; 11261 } 11262 } 11263 } 11264 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11265 throw e.setUnfinishedMessage(this); 11266 } catch (java.io.IOException e) { 11267 throw new com.google.protobuf.InvalidProtocolBufferException( 11268 e.getMessage()).setUnfinishedMessage(this); 11269 } finally { 11270 this.unknownFields = unknownFields.build(); 11271 makeExtensionsImmutable(); 11272 } 11273 } 11274 public static final com.google.protobuf.Descriptors.Descriptor 11275 getDescriptor() { 11276 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor; 11277 } 11278 11279 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11280 internalGetFieldAccessorTable() { 11281 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable 11282 .ensureFieldAccessorsInitialized( 11283 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class); 11284 } 11285 11286 public static com.google.protobuf.Parser<FormatResponseProto> PARSER = 11287 new com.google.protobuf.AbstractParser<FormatResponseProto>() { 11288 public FormatResponseProto parsePartialFrom( 11289 com.google.protobuf.CodedInputStream input, 11290 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11291 throws com.google.protobuf.InvalidProtocolBufferException { 11292 return new FormatResponseProto(input, extensionRegistry); 11293 } 11294 }; 11295 11296 @java.lang.Override 11297 public com.google.protobuf.Parser<FormatResponseProto> getParserForType() { 11298 return PARSER; 11299 } 11300 11301 private void initFields() { 11302 } 11303 private byte memoizedIsInitialized = -1; 11304 public final boolean isInitialized() { 11305 byte isInitialized = memoizedIsInitialized; 11306 if (isInitialized != -1) return isInitialized == 1; 11307 11308 memoizedIsInitialized = 1; 11309 return true; 11310 } 11311 11312 public void writeTo(com.google.protobuf.CodedOutputStream output) 11313 throws java.io.IOException { 11314 getSerializedSize(); 11315 getUnknownFields().writeTo(output); 11316 } 11317 11318 private int memoizedSerializedSize = -1; 11319 public int getSerializedSize() { 11320 int size = memoizedSerializedSize; 11321 if (size != -1) return size; 11322 11323 size = 0; 11324 size += getUnknownFields().getSerializedSize(); 11325 memoizedSerializedSize = size; 11326 return size; 11327 } 11328 11329 private static final long serialVersionUID = 0L; 11330 @java.lang.Override 11331 protected java.lang.Object writeReplace() 11332 throws java.io.ObjectStreamException { 11333 return super.writeReplace(); 11334 } 11335 11336 @java.lang.Override 11337 public boolean equals(final java.lang.Object obj) { 11338 if (obj == this) { 11339 return true; 11340 } 11341 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)) { 11342 return super.equals(obj); 11343 } 11344 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) obj; 11345 11346 boolean result = true; 11347 result = result && 11348 getUnknownFields().equals(other.getUnknownFields()); 11349 return result; 11350 } 11351 11352 private int memoizedHashCode = 0; 11353 @java.lang.Override 11354 public int hashCode() { 11355 if (memoizedHashCode != 0) { 11356 return memoizedHashCode; 11357 } 11358 int hash = 41; 11359 hash = (19 * hash) + getDescriptorForType().hashCode(); 11360 hash = (29 * hash) + getUnknownFields().hashCode(); 11361 memoizedHashCode = hash; 11362 return hash; 11363 } 11364 11365 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11366 com.google.protobuf.ByteString data) 11367 throws com.google.protobuf.InvalidProtocolBufferException { 11368 return PARSER.parseFrom(data); 11369 } 11370 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11371 com.google.protobuf.ByteString data, 11372 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11373 throws com.google.protobuf.InvalidProtocolBufferException { 11374 return PARSER.parseFrom(data, extensionRegistry); 11375 } 11376 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(byte[] data) 11377 throws com.google.protobuf.InvalidProtocolBufferException { 11378 return PARSER.parseFrom(data); 11379 } 11380 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11381 byte[] data, 11382 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11383 throws com.google.protobuf.InvalidProtocolBufferException { 11384 return PARSER.parseFrom(data, extensionRegistry); 11385 } 11386 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(java.io.InputStream input) 11387 throws java.io.IOException { 11388 return PARSER.parseFrom(input); 11389 } 11390 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11391 java.io.InputStream input, 11392 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11393 throws java.io.IOException { 11394 return PARSER.parseFrom(input, extensionRegistry); 11395 } 11396 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(java.io.InputStream input) 11397 throws java.io.IOException { 11398 return PARSER.parseDelimitedFrom(input); 11399 } 11400 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom( 11401 java.io.InputStream input, 11402 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11403 throws java.io.IOException { 11404 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11405 } 11406 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11407 com.google.protobuf.CodedInputStream input) 11408 throws java.io.IOException { 11409 return PARSER.parseFrom(input); 11410 } 11411 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom( 11412 com.google.protobuf.CodedInputStream input, 11413 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11414 throws java.io.IOException { 11415 return PARSER.parseFrom(input, extensionRegistry); 11416 } 11417 11418 public static Builder newBuilder() { return Builder.create(); } 11419 public Builder newBuilderForType() { return newBuilder(); } 11420 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto prototype) { 11421 return newBuilder().mergeFrom(prototype); 11422 } 11423 public Builder toBuilder() { return newBuilder(this); } 11424 11425 @java.lang.Override 11426 protected Builder newBuilderForType( 11427 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11428 Builder builder = new Builder(parent); 11429 return builder; 11430 } 11431 /** 11432 * Protobuf type {@code hadoop.hdfs.FormatResponseProto} 11433 */ 11434 public static final class Builder extends 11435 com.google.protobuf.GeneratedMessage.Builder<Builder> 11436 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProtoOrBuilder { 11437 public static final com.google.protobuf.Descriptors.Descriptor 11438 getDescriptor() { 11439 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor; 11440 } 11441 11442 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11443 internalGetFieldAccessorTable() { 11444 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable 11445 .ensureFieldAccessorsInitialized( 11446 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class); 11447 } 11448 11449 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.newBuilder() 11450 private Builder() { 11451 maybeForceBuilderInitialization(); 11452 } 11453 11454 private Builder( 11455 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11456 super(parent); 11457 maybeForceBuilderInitialization(); 11458 } 11459 private void maybeForceBuilderInitialization() { 11460 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 11461 } 11462 } 11463 private static Builder create() { 11464 return new Builder(); 11465 } 11466 11467 public Builder clear() { 11468 super.clear(); 11469 return this; 11470 } 11471 11472 public Builder clone() { 11473 return create().mergeFrom(buildPartial()); 11474 } 11475 11476 public com.google.protobuf.Descriptors.Descriptor 11477 getDescriptorForType() { 11478 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor; 11479 } 11480 11481 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto getDefaultInstanceForType() { 11482 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(); 11483 } 11484 11485 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto build() { 11486 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = buildPartial(); 11487 if (!result.isInitialized()) { 11488 throw newUninitializedMessageException(result); 11489 } 11490 return result; 11491 } 11492 11493 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto buildPartial() { 11494 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto(this); 11495 onBuilt(); 11496 return result; 11497 } 11498 11499 public Builder mergeFrom(com.google.protobuf.Message other) { 11500 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) { 11501 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)other); 11502 } else { 11503 super.mergeFrom(other); 11504 return this; 11505 } 11506 } 11507 11508 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other) { 11509 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()) return this; 11510 this.mergeUnknownFields(other.getUnknownFields()); 11511 return this; 11512 } 11513 11514 public final boolean isInitialized() { 11515 return true; 11516 } 11517 11518 public Builder mergeFrom( 11519 com.google.protobuf.CodedInputStream input, 11520 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11521 throws java.io.IOException { 11522 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parsedMessage = null; 11523 try { 11524 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 11525 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11526 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) e.getUnfinishedMessage(); 11527 throw e; 11528 } finally { 11529 if (parsedMessage != null) { 11530 mergeFrom(parsedMessage); 11531 } 11532 } 11533 return this; 11534 } 11535 11536 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatResponseProto) 11537 } 11538 11539 static { 11540 defaultInstance = new FormatResponseProto(true); 11541 defaultInstance.initFields(); 11542 } 11543 11544 // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatResponseProto) 11545 } 11546 11547 public interface NewEpochRequestProtoOrBuilder 11548 extends com.google.protobuf.MessageOrBuilder { 11549 11550 // required .hadoop.hdfs.JournalIdProto jid = 1; 11551 /** 11552 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11553 */ 11554 boolean hasJid(); 11555 /** 11556 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11557 */ 11558 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid(); 11559 /** 11560 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11561 */ 11562 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder(); 11563 11564 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 11565 /** 11566 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11567 */ 11568 boolean hasNsInfo(); 11569 /** 11570 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11571 */ 11572 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo(); 11573 /** 11574 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11575 */ 11576 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder(); 11577 11578 // required uint64 epoch = 3; 11579 /** 11580 * <code>required uint64 epoch = 3;</code> 11581 */ 11582 boolean hasEpoch(); 11583 /** 11584 * <code>required uint64 epoch = 3;</code> 11585 */ 11586 long getEpoch(); 11587 } 11588 /** 11589 * Protobuf type {@code hadoop.hdfs.NewEpochRequestProto} 11590 * 11591 * <pre> 11592 ** 11593 * newEpoch() 11594 * </pre> 11595 */ 11596 public static final class NewEpochRequestProto extends 11597 com.google.protobuf.GeneratedMessage 11598 implements NewEpochRequestProtoOrBuilder { 11599 // Use NewEpochRequestProto.newBuilder() to construct. 11600 private NewEpochRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 11601 super(builder); 11602 this.unknownFields = builder.getUnknownFields(); 11603 } 11604 private NewEpochRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 11605 11606 private static final NewEpochRequestProto defaultInstance; 11607 public static NewEpochRequestProto getDefaultInstance() { 11608 return defaultInstance; 11609 } 11610 11611 public NewEpochRequestProto getDefaultInstanceForType() { 11612 return defaultInstance; 11613 } 11614 11615 private final com.google.protobuf.UnknownFieldSet unknownFields; 11616 @java.lang.Override 11617 public final com.google.protobuf.UnknownFieldSet 11618 getUnknownFields() { 11619 return this.unknownFields; 11620 } 11621 private NewEpochRequestProto( 11622 com.google.protobuf.CodedInputStream input, 11623 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11624 throws com.google.protobuf.InvalidProtocolBufferException { 11625 initFields(); 11626 int mutable_bitField0_ = 0; 11627 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 11628 com.google.protobuf.UnknownFieldSet.newBuilder(); 11629 try { 11630 boolean done = false; 11631 while (!done) { 11632 int tag = input.readTag(); 11633 switch (tag) { 11634 case 0: 11635 done = true; 11636 break; 11637 default: { 11638 if (!parseUnknownField(input, unknownFields, 11639 extensionRegistry, tag)) { 11640 done = true; 11641 } 11642 break; 11643 } 11644 case 10: { 11645 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 11646 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11647 subBuilder = jid_.toBuilder(); 11648 } 11649 jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 11650 if (subBuilder != null) { 11651 subBuilder.mergeFrom(jid_); 11652 jid_ = subBuilder.buildPartial(); 11653 } 11654 bitField0_ |= 0x00000001; 11655 break; 11656 } 11657 case 18: { 11658 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = null; 11659 if (((bitField0_ & 0x00000002) == 0x00000002)) { 11660 subBuilder = nsInfo_.toBuilder(); 11661 } 11662 nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.PARSER, extensionRegistry); 11663 if (subBuilder != null) { 11664 subBuilder.mergeFrom(nsInfo_); 11665 nsInfo_ = subBuilder.buildPartial(); 11666 } 11667 bitField0_ |= 0x00000002; 11668 break; 11669 } 11670 case 24: { 11671 bitField0_ |= 0x00000004; 11672 epoch_ = input.readUInt64(); 11673 break; 11674 } 11675 } 11676 } 11677 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11678 throw e.setUnfinishedMessage(this); 11679 } catch (java.io.IOException e) { 11680 throw new com.google.protobuf.InvalidProtocolBufferException( 11681 e.getMessage()).setUnfinishedMessage(this); 11682 } finally { 11683 this.unknownFields = unknownFields.build(); 11684 makeExtensionsImmutable(); 11685 } 11686 } 11687 public static final com.google.protobuf.Descriptors.Descriptor 11688 getDescriptor() { 11689 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor; 11690 } 11691 11692 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11693 internalGetFieldAccessorTable() { 11694 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable 11695 .ensureFieldAccessorsInitialized( 11696 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class); 11697 } 11698 11699 public static com.google.protobuf.Parser<NewEpochRequestProto> PARSER = 11700 new com.google.protobuf.AbstractParser<NewEpochRequestProto>() { 11701 public NewEpochRequestProto parsePartialFrom( 11702 com.google.protobuf.CodedInputStream input, 11703 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11704 throws com.google.protobuf.InvalidProtocolBufferException { 11705 return new NewEpochRequestProto(input, extensionRegistry); 11706 } 11707 }; 11708 11709 @java.lang.Override 11710 public com.google.protobuf.Parser<NewEpochRequestProto> getParserForType() { 11711 return PARSER; 11712 } 11713 11714 private int bitField0_; 11715 // required .hadoop.hdfs.JournalIdProto jid = 1; 11716 public static final int JID_FIELD_NUMBER = 1; 11717 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_; 11718 /** 11719 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11720 */ 11721 public boolean hasJid() { 11722 return ((bitField0_ & 0x00000001) == 0x00000001); 11723 } 11724 /** 11725 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11726 */ 11727 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 11728 return jid_; 11729 } 11730 /** 11731 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 11732 */ 11733 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 11734 return jid_; 11735 } 11736 11737 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 11738 public static final int NSINFO_FIELD_NUMBER = 2; 11739 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_; 11740 /** 11741 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11742 */ 11743 public boolean hasNsInfo() { 11744 return ((bitField0_ & 0x00000002) == 0x00000002); 11745 } 11746 /** 11747 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11748 */ 11749 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() { 11750 return nsInfo_; 11751 } 11752 /** 11753 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 11754 */ 11755 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() { 11756 return nsInfo_; 11757 } 11758 11759 // required uint64 epoch = 3; 11760 public static final int EPOCH_FIELD_NUMBER = 3; 11761 private long epoch_; 11762 /** 11763 * <code>required uint64 epoch = 3;</code> 11764 */ 11765 public boolean hasEpoch() { 11766 return ((bitField0_ & 0x00000004) == 0x00000004); 11767 } 11768 /** 11769 * <code>required uint64 epoch = 3;</code> 11770 */ 11771 public long getEpoch() { 11772 return epoch_; 11773 } 11774 11775 private void initFields() { 11776 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 11777 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 11778 epoch_ = 0L; 11779 } 11780 private byte memoizedIsInitialized = -1; 11781 public final boolean isInitialized() { 11782 byte isInitialized = memoizedIsInitialized; 11783 if (isInitialized != -1) return isInitialized == 1; 11784 11785 if (!hasJid()) { 11786 memoizedIsInitialized = 0; 11787 return false; 11788 } 11789 if (!hasNsInfo()) { 11790 memoizedIsInitialized = 0; 11791 return false; 11792 } 11793 if (!hasEpoch()) { 11794 memoizedIsInitialized = 0; 11795 return false; 11796 } 11797 if (!getJid().isInitialized()) { 11798 memoizedIsInitialized = 0; 11799 return false; 11800 } 11801 if (!getNsInfo().isInitialized()) { 11802 memoizedIsInitialized = 0; 11803 return false; 11804 } 11805 memoizedIsInitialized = 1; 11806 return true; 11807 } 11808 11809 public void writeTo(com.google.protobuf.CodedOutputStream output) 11810 throws java.io.IOException { 11811 getSerializedSize(); 11812 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11813 output.writeMessage(1, jid_); 11814 } 11815 if (((bitField0_ & 0x00000002) == 0x00000002)) { 11816 output.writeMessage(2, nsInfo_); 11817 } 11818 if (((bitField0_ & 0x00000004) == 0x00000004)) { 11819 output.writeUInt64(3, epoch_); 11820 } 11821 getUnknownFields().writeTo(output); 11822 } 11823 11824 private int memoizedSerializedSize = -1; 11825 public int getSerializedSize() { 11826 int size = memoizedSerializedSize; 11827 if (size != -1) return size; 11828 11829 size = 0; 11830 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11831 size += com.google.protobuf.CodedOutputStream 11832 .computeMessageSize(1, jid_); 11833 } 11834 if (((bitField0_ & 0x00000002) == 0x00000002)) { 11835 size += com.google.protobuf.CodedOutputStream 11836 .computeMessageSize(2, nsInfo_); 11837 } 11838 if (((bitField0_ & 0x00000004) == 0x00000004)) { 11839 size += com.google.protobuf.CodedOutputStream 11840 .computeUInt64Size(3, epoch_); 11841 } 11842 size += getUnknownFields().getSerializedSize(); 11843 memoizedSerializedSize = size; 11844 return size; 11845 } 11846 11847 private static final long serialVersionUID = 0L; 11848 @java.lang.Override 11849 protected java.lang.Object writeReplace() 11850 throws java.io.ObjectStreamException { 11851 return super.writeReplace(); 11852 } 11853 11854 @java.lang.Override 11855 public boolean equals(final java.lang.Object obj) { 11856 if (obj == this) { 11857 return true; 11858 } 11859 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)) { 11860 return super.equals(obj); 11861 } 11862 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) obj; 11863 11864 boolean result = true; 11865 result = result && (hasJid() == other.hasJid()); 11866 if (hasJid()) { 11867 result = result && getJid() 11868 .equals(other.getJid()); 11869 } 11870 result = result && (hasNsInfo() == other.hasNsInfo()); 11871 if (hasNsInfo()) { 11872 result = result && getNsInfo() 11873 .equals(other.getNsInfo()); 11874 } 11875 result = result && (hasEpoch() == other.hasEpoch()); 11876 if (hasEpoch()) { 11877 result = result && (getEpoch() 11878 == other.getEpoch()); 11879 } 11880 result = result && 11881 getUnknownFields().equals(other.getUnknownFields()); 11882 return result; 11883 } 11884 11885 private int memoizedHashCode = 0; 11886 @java.lang.Override 11887 public int hashCode() { 11888 if (memoizedHashCode != 0) { 11889 return memoizedHashCode; 11890 } 11891 int hash = 41; 11892 hash = (19 * hash) + getDescriptorForType().hashCode(); 11893 if (hasJid()) { 11894 hash = (37 * hash) + JID_FIELD_NUMBER; 11895 hash = (53 * hash) + getJid().hashCode(); 11896 } 11897 if (hasNsInfo()) { 11898 hash = (37 * hash) + NSINFO_FIELD_NUMBER; 11899 hash = (53 * hash) + getNsInfo().hashCode(); 11900 } 11901 if (hasEpoch()) { 11902 hash = (37 * hash) + EPOCH_FIELD_NUMBER; 11903 hash = (53 * hash) + hashLong(getEpoch()); 11904 } 11905 hash = (29 * hash) + getUnknownFields().hashCode(); 11906 memoizedHashCode = hash; 11907 return hash; 11908 } 11909 11910 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11911 com.google.protobuf.ByteString data) 11912 throws com.google.protobuf.InvalidProtocolBufferException { 11913 return PARSER.parseFrom(data); 11914 } 11915 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11916 com.google.protobuf.ByteString data, 11917 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11918 throws com.google.protobuf.InvalidProtocolBufferException { 11919 return PARSER.parseFrom(data, extensionRegistry); 11920 } 11921 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(byte[] data) 11922 throws com.google.protobuf.InvalidProtocolBufferException { 11923 return PARSER.parseFrom(data); 11924 } 11925 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11926 byte[] data, 11927 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11928 throws com.google.protobuf.InvalidProtocolBufferException { 11929 return PARSER.parseFrom(data, extensionRegistry); 11930 } 11931 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(java.io.InputStream input) 11932 throws java.io.IOException { 11933 return PARSER.parseFrom(input); 11934 } 11935 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11936 java.io.InputStream input, 11937 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11938 throws java.io.IOException { 11939 return PARSER.parseFrom(input, extensionRegistry); 11940 } 11941 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(java.io.InputStream input) 11942 throws java.io.IOException { 11943 return PARSER.parseDelimitedFrom(input); 11944 } 11945 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom( 11946 java.io.InputStream input, 11947 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11948 throws java.io.IOException { 11949 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11950 } 11951 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11952 com.google.protobuf.CodedInputStream input) 11953 throws java.io.IOException { 11954 return PARSER.parseFrom(input); 11955 } 11956 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom( 11957 com.google.protobuf.CodedInputStream input, 11958 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11959 throws java.io.IOException { 11960 return PARSER.parseFrom(input, extensionRegistry); 11961 } 11962 11963 public static Builder newBuilder() { return Builder.create(); } 11964 public Builder newBuilderForType() { return newBuilder(); } 11965 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto prototype) { 11966 return newBuilder().mergeFrom(prototype); 11967 } 11968 public Builder toBuilder() { return newBuilder(this); } 11969 11970 @java.lang.Override 11971 protected Builder newBuilderForType( 11972 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11973 Builder builder = new Builder(parent); 11974 return builder; 11975 } 11976 /** 11977 * Protobuf type {@code hadoop.hdfs.NewEpochRequestProto} 11978 * 11979 * <pre> 11980 ** 11981 * newEpoch() 11982 * </pre> 11983 */ 11984 public static final class Builder extends 11985 com.google.protobuf.GeneratedMessage.Builder<Builder> 11986 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProtoOrBuilder { 11987 public static final com.google.protobuf.Descriptors.Descriptor 11988 getDescriptor() { 11989 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor; 11990 } 11991 11992 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11993 internalGetFieldAccessorTable() { 11994 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable 11995 .ensureFieldAccessorsInitialized( 11996 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class); 11997 } 11998 11999 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.newBuilder() 12000 private Builder() { 12001 maybeForceBuilderInitialization(); 12002 } 12003 12004 private Builder( 12005 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12006 super(parent); 12007 maybeForceBuilderInitialization(); 12008 } 12009 private void maybeForceBuilderInitialization() { 12010 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 12011 getJidFieldBuilder(); 12012 getNsInfoFieldBuilder(); 12013 } 12014 } 12015 private static Builder create() { 12016 return new Builder(); 12017 } 12018 12019 public Builder clear() { 12020 super.clear(); 12021 if (jidBuilder_ == null) { 12022 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 12023 } else { 12024 jidBuilder_.clear(); 12025 } 12026 bitField0_ = (bitField0_ & ~0x00000001); 12027 if (nsInfoBuilder_ == null) { 12028 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 12029 } else { 12030 nsInfoBuilder_.clear(); 12031 } 12032 bitField0_ = (bitField0_ & ~0x00000002); 12033 epoch_ = 0L; 12034 bitField0_ = (bitField0_ & ~0x00000004); 12035 return this; 12036 } 12037 12038 public Builder clone() { 12039 return create().mergeFrom(buildPartial()); 12040 } 12041 12042 public com.google.protobuf.Descriptors.Descriptor 12043 getDescriptorForType() { 12044 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor; 12045 } 12046 12047 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto getDefaultInstanceForType() { 12048 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance(); 12049 } 12050 12051 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto build() { 12052 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = buildPartial(); 12053 if (!result.isInitialized()) { 12054 throw newUninitializedMessageException(result); 12055 } 12056 return result; 12057 } 12058 12059 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto buildPartial() { 12060 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto(this); 12061 int from_bitField0_ = bitField0_; 12062 int to_bitField0_ = 0; 12063 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 12064 to_bitField0_ |= 0x00000001; 12065 } 12066 if (jidBuilder_ == null) { 12067 result.jid_ = jid_; 12068 } else { 12069 result.jid_ = jidBuilder_.build(); 12070 } 12071 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 12072 to_bitField0_ |= 0x00000002; 12073 } 12074 if (nsInfoBuilder_ == null) { 12075 result.nsInfo_ = nsInfo_; 12076 } else { 12077 result.nsInfo_ = nsInfoBuilder_.build(); 12078 } 12079 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 12080 to_bitField0_ |= 0x00000004; 12081 } 12082 result.epoch_ = epoch_; 12083 result.bitField0_ = to_bitField0_; 12084 onBuilt(); 12085 return result; 12086 } 12087 12088 public Builder mergeFrom(com.google.protobuf.Message other) { 12089 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) { 12090 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)other); 12091 } else { 12092 super.mergeFrom(other); 12093 return this; 12094 } 12095 } 12096 12097 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other) { 12098 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance()) return this; 12099 if (other.hasJid()) { 12100 mergeJid(other.getJid()); 12101 } 12102 if (other.hasNsInfo()) { 12103 mergeNsInfo(other.getNsInfo()); 12104 } 12105 if (other.hasEpoch()) { 12106 setEpoch(other.getEpoch()); 12107 } 12108 this.mergeUnknownFields(other.getUnknownFields()); 12109 return this; 12110 } 12111 12112 public final boolean isInitialized() { 12113 if (!hasJid()) { 12114 12115 return false; 12116 } 12117 if (!hasNsInfo()) { 12118 12119 return false; 12120 } 12121 if (!hasEpoch()) { 12122 12123 return false; 12124 } 12125 if (!getJid().isInitialized()) { 12126 12127 return false; 12128 } 12129 if (!getNsInfo().isInitialized()) { 12130 12131 return false; 12132 } 12133 return true; 12134 } 12135 12136 public Builder mergeFrom( 12137 com.google.protobuf.CodedInputStream input, 12138 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12139 throws java.io.IOException { 12140 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parsedMessage = null; 12141 try { 12142 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 12143 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12144 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) e.getUnfinishedMessage(); 12145 throw e; 12146 } finally { 12147 if (parsedMessage != null) { 12148 mergeFrom(parsedMessage); 12149 } 12150 } 12151 return this; 12152 } 12153 private int bitField0_; 12154 12155 // required .hadoop.hdfs.JournalIdProto jid = 1; 12156 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 12157 private com.google.protobuf.SingleFieldBuilder< 12158 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_; 12159 /** 12160 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12161 */ 12162 public boolean hasJid() { 12163 return ((bitField0_ & 0x00000001) == 0x00000001); 12164 } 12165 /** 12166 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12167 */ 12168 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 12169 if (jidBuilder_ == null) { 12170 return jid_; 12171 } else { 12172 return jidBuilder_.getMessage(); 12173 } 12174 } 12175 /** 12176 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12177 */ 12178 public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 12179 if (jidBuilder_ == null) { 12180 if (value == null) { 12181 throw new NullPointerException(); 12182 } 12183 jid_ = value; 12184 onChanged(); 12185 } else { 12186 jidBuilder_.setMessage(value); 12187 } 12188 bitField0_ |= 0x00000001; 12189 return this; 12190 } 12191 /** 12192 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12193 */ 12194 public Builder setJid( 12195 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 12196 if (jidBuilder_ == null) { 12197 jid_ = builderForValue.build(); 12198 onChanged(); 12199 } else { 12200 jidBuilder_.setMessage(builderForValue.build()); 12201 } 12202 bitField0_ |= 0x00000001; 12203 return this; 12204 } 12205 /** 12206 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12207 */ 12208 public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 12209 if (jidBuilder_ == null) { 12210 if (((bitField0_ & 0x00000001) == 0x00000001) && 12211 jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 12212 jid_ = 12213 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial(); 12214 } else { 12215 jid_ = value; 12216 } 12217 onChanged(); 12218 } else { 12219 jidBuilder_.mergeFrom(value); 12220 } 12221 bitField0_ |= 0x00000001; 12222 return this; 12223 } 12224 /** 12225 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12226 */ 12227 public Builder clearJid() { 12228 if (jidBuilder_ == null) { 12229 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 12230 onChanged(); 12231 } else { 12232 jidBuilder_.clear(); 12233 } 12234 bitField0_ = (bitField0_ & ~0x00000001); 12235 return this; 12236 } 12237 /** 12238 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12239 */ 12240 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() { 12241 bitField0_ |= 0x00000001; 12242 onChanged(); 12243 return getJidFieldBuilder().getBuilder(); 12244 } 12245 /** 12246 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12247 */ 12248 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 12249 if (jidBuilder_ != null) { 12250 return jidBuilder_.getMessageOrBuilder(); 12251 } else { 12252 return jid_; 12253 } 12254 } 12255 /** 12256 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12257 */ 12258 private com.google.protobuf.SingleFieldBuilder< 12259 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 12260 getJidFieldBuilder() { 12261 if (jidBuilder_ == null) { 12262 jidBuilder_ = new com.google.protobuf.SingleFieldBuilder< 12263 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 12264 jid_, 12265 getParentForChildren(), 12266 isClean()); 12267 jid_ = null; 12268 } 12269 return jidBuilder_; 12270 } 12271 12272 // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2; 12273 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 12274 private com.google.protobuf.SingleFieldBuilder< 12275 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_; 12276 /** 12277 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12278 */ 12279 public boolean hasNsInfo() { 12280 return ((bitField0_ & 0x00000002) == 0x00000002); 12281 } 12282 /** 12283 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12284 */ 12285 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() { 12286 if (nsInfoBuilder_ == null) { 12287 return nsInfo_; 12288 } else { 12289 return nsInfoBuilder_.getMessage(); 12290 } 12291 } 12292 /** 12293 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12294 */ 12295 public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) { 12296 if (nsInfoBuilder_ == null) { 12297 if (value == null) { 12298 throw new NullPointerException(); 12299 } 12300 nsInfo_ = value; 12301 onChanged(); 12302 } else { 12303 nsInfoBuilder_.setMessage(value); 12304 } 12305 bitField0_ |= 0x00000002; 12306 return this; 12307 } 12308 /** 12309 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12310 */ 12311 public Builder setNsInfo( 12312 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) { 12313 if (nsInfoBuilder_ == null) { 12314 nsInfo_ = builderForValue.build(); 12315 onChanged(); 12316 } else { 12317 nsInfoBuilder_.setMessage(builderForValue.build()); 12318 } 12319 bitField0_ |= 0x00000002; 12320 return this; 12321 } 12322 /** 12323 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12324 */ 12325 public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) { 12326 if (nsInfoBuilder_ == null) { 12327 if (((bitField0_ & 0x00000002) == 0x00000002) && 12328 nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) { 12329 nsInfo_ = 12330 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial(); 12331 } else { 12332 nsInfo_ = value; 12333 } 12334 onChanged(); 12335 } else { 12336 nsInfoBuilder_.mergeFrom(value); 12337 } 12338 bitField0_ |= 0x00000002; 12339 return this; 12340 } 12341 /** 12342 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12343 */ 12344 public Builder clearNsInfo() { 12345 if (nsInfoBuilder_ == null) { 12346 nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance(); 12347 onChanged(); 12348 } else { 12349 nsInfoBuilder_.clear(); 12350 } 12351 bitField0_ = (bitField0_ & ~0x00000002); 12352 return this; 12353 } 12354 /** 12355 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12356 */ 12357 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() { 12358 bitField0_ |= 0x00000002; 12359 onChanged(); 12360 return getNsInfoFieldBuilder().getBuilder(); 12361 } 12362 /** 12363 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12364 */ 12365 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() { 12366 if (nsInfoBuilder_ != null) { 12367 return nsInfoBuilder_.getMessageOrBuilder(); 12368 } else { 12369 return nsInfo_; 12370 } 12371 } 12372 /** 12373 * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code> 12374 */ 12375 private com.google.protobuf.SingleFieldBuilder< 12376 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 12377 getNsInfoFieldBuilder() { 12378 if (nsInfoBuilder_ == null) { 12379 nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 12380 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>( 12381 nsInfo_, 12382 getParentForChildren(), 12383 isClean()); 12384 nsInfo_ = null; 12385 } 12386 return nsInfoBuilder_; 12387 } 12388 12389 // required uint64 epoch = 3; 12390 private long epoch_ ; 12391 /** 12392 * <code>required uint64 epoch = 3;</code> 12393 */ 12394 public boolean hasEpoch() { 12395 return ((bitField0_ & 0x00000004) == 0x00000004); 12396 } 12397 /** 12398 * <code>required uint64 epoch = 3;</code> 12399 */ 12400 public long getEpoch() { 12401 return epoch_; 12402 } 12403 /** 12404 * <code>required uint64 epoch = 3;</code> 12405 */ 12406 public Builder setEpoch(long value) { 12407 bitField0_ |= 0x00000004; 12408 epoch_ = value; 12409 onChanged(); 12410 return this; 12411 } 12412 /** 12413 * <code>required uint64 epoch = 3;</code> 12414 */ 12415 public Builder clearEpoch() { 12416 bitField0_ = (bitField0_ & ~0x00000004); 12417 epoch_ = 0L; 12418 onChanged(); 12419 return this; 12420 } 12421 12422 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochRequestProto) 12423 } 12424 12425 static { 12426 defaultInstance = new NewEpochRequestProto(true); 12427 defaultInstance.initFields(); 12428 } 12429 12430 // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochRequestProto) 12431 } 12432 12433 public interface NewEpochResponseProtoOrBuilder 12434 extends com.google.protobuf.MessageOrBuilder { 12435 12436 // optional uint64 lastSegmentTxId = 1; 12437 /** 12438 * <code>optional uint64 lastSegmentTxId = 1;</code> 12439 */ 12440 boolean hasLastSegmentTxId(); 12441 /** 12442 * <code>optional uint64 lastSegmentTxId = 1;</code> 12443 */ 12444 long getLastSegmentTxId(); 12445 } 12446 /** 12447 * Protobuf type {@code hadoop.hdfs.NewEpochResponseProto} 12448 */ 12449 public static final class NewEpochResponseProto extends 12450 com.google.protobuf.GeneratedMessage 12451 implements NewEpochResponseProtoOrBuilder { 12452 // Use NewEpochResponseProto.newBuilder() to construct. 12453 private NewEpochResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 12454 super(builder); 12455 this.unknownFields = builder.getUnknownFields(); 12456 } 12457 private NewEpochResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 12458 12459 private static final NewEpochResponseProto defaultInstance; 12460 public static NewEpochResponseProto getDefaultInstance() { 12461 return defaultInstance; 12462 } 12463 12464 public NewEpochResponseProto getDefaultInstanceForType() { 12465 return defaultInstance; 12466 } 12467 12468 private final com.google.protobuf.UnknownFieldSet unknownFields; 12469 @java.lang.Override 12470 public final com.google.protobuf.UnknownFieldSet 12471 getUnknownFields() { 12472 return this.unknownFields; 12473 } 12474 private NewEpochResponseProto( 12475 com.google.protobuf.CodedInputStream input, 12476 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12477 throws com.google.protobuf.InvalidProtocolBufferException { 12478 initFields(); 12479 int mutable_bitField0_ = 0; 12480 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 12481 com.google.protobuf.UnknownFieldSet.newBuilder(); 12482 try { 12483 boolean done = false; 12484 while (!done) { 12485 int tag = input.readTag(); 12486 switch (tag) { 12487 case 0: 12488 done = true; 12489 break; 12490 default: { 12491 if (!parseUnknownField(input, unknownFields, 12492 extensionRegistry, tag)) { 12493 done = true; 12494 } 12495 break; 12496 } 12497 case 8: { 12498 bitField0_ |= 0x00000001; 12499 lastSegmentTxId_ = input.readUInt64(); 12500 break; 12501 } 12502 } 12503 } 12504 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12505 throw e.setUnfinishedMessage(this); 12506 } catch (java.io.IOException e) { 12507 throw new com.google.protobuf.InvalidProtocolBufferException( 12508 e.getMessage()).setUnfinishedMessage(this); 12509 } finally { 12510 this.unknownFields = unknownFields.build(); 12511 makeExtensionsImmutable(); 12512 } 12513 } 12514 public static final com.google.protobuf.Descriptors.Descriptor 12515 getDescriptor() { 12516 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor; 12517 } 12518 12519 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12520 internalGetFieldAccessorTable() { 12521 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable 12522 .ensureFieldAccessorsInitialized( 12523 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class); 12524 } 12525 12526 public static com.google.protobuf.Parser<NewEpochResponseProto> PARSER = 12527 new com.google.protobuf.AbstractParser<NewEpochResponseProto>() { 12528 public NewEpochResponseProto parsePartialFrom( 12529 com.google.protobuf.CodedInputStream input, 12530 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12531 throws com.google.protobuf.InvalidProtocolBufferException { 12532 return new NewEpochResponseProto(input, extensionRegistry); 12533 } 12534 }; 12535 12536 @java.lang.Override 12537 public com.google.protobuf.Parser<NewEpochResponseProto> getParserForType() { 12538 return PARSER; 12539 } 12540 12541 private int bitField0_; 12542 // optional uint64 lastSegmentTxId = 1; 12543 public static final int LASTSEGMENTTXID_FIELD_NUMBER = 1; 12544 private long lastSegmentTxId_; 12545 /** 12546 * <code>optional uint64 lastSegmentTxId = 1;</code> 12547 */ 12548 public boolean hasLastSegmentTxId() { 12549 return ((bitField0_ & 0x00000001) == 0x00000001); 12550 } 12551 /** 12552 * <code>optional uint64 lastSegmentTxId = 1;</code> 12553 */ 12554 public long getLastSegmentTxId() { 12555 return lastSegmentTxId_; 12556 } 12557 12558 private void initFields() { 12559 lastSegmentTxId_ = 0L; 12560 } 12561 private byte memoizedIsInitialized = -1; 12562 public final boolean isInitialized() { 12563 byte isInitialized = memoizedIsInitialized; 12564 if (isInitialized != -1) return isInitialized == 1; 12565 12566 memoizedIsInitialized = 1; 12567 return true; 12568 } 12569 12570 public void writeTo(com.google.protobuf.CodedOutputStream output) 12571 throws java.io.IOException { 12572 getSerializedSize(); 12573 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12574 output.writeUInt64(1, lastSegmentTxId_); 12575 } 12576 getUnknownFields().writeTo(output); 12577 } 12578 12579 private int memoizedSerializedSize = -1; 12580 public int getSerializedSize() { 12581 int size = memoizedSerializedSize; 12582 if (size != -1) return size; 12583 12584 size = 0; 12585 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12586 size += com.google.protobuf.CodedOutputStream 12587 .computeUInt64Size(1, lastSegmentTxId_); 12588 } 12589 size += getUnknownFields().getSerializedSize(); 12590 memoizedSerializedSize = size; 12591 return size; 12592 } 12593 12594 private static final long serialVersionUID = 0L; 12595 @java.lang.Override 12596 protected java.lang.Object writeReplace() 12597 throws java.io.ObjectStreamException { 12598 return super.writeReplace(); 12599 } 12600 12601 @java.lang.Override 12602 public boolean equals(final java.lang.Object obj) { 12603 if (obj == this) { 12604 return true; 12605 } 12606 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)) { 12607 return super.equals(obj); 12608 } 12609 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) obj; 12610 12611 boolean result = true; 12612 result = result && (hasLastSegmentTxId() == other.hasLastSegmentTxId()); 12613 if (hasLastSegmentTxId()) { 12614 result = result && (getLastSegmentTxId() 12615 == other.getLastSegmentTxId()); 12616 } 12617 result = result && 12618 getUnknownFields().equals(other.getUnknownFields()); 12619 return result; 12620 } 12621 12622 private int memoizedHashCode = 0; 12623 @java.lang.Override 12624 public int hashCode() { 12625 if (memoizedHashCode != 0) { 12626 return memoizedHashCode; 12627 } 12628 int hash = 41; 12629 hash = (19 * hash) + getDescriptorForType().hashCode(); 12630 if (hasLastSegmentTxId()) { 12631 hash = (37 * hash) + LASTSEGMENTTXID_FIELD_NUMBER; 12632 hash = (53 * hash) + hashLong(getLastSegmentTxId()); 12633 } 12634 hash = (29 * hash) + getUnknownFields().hashCode(); 12635 memoizedHashCode = hash; 12636 return hash; 12637 } 12638 12639 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12640 com.google.protobuf.ByteString data) 12641 throws com.google.protobuf.InvalidProtocolBufferException { 12642 return PARSER.parseFrom(data); 12643 } 12644 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12645 com.google.protobuf.ByteString data, 12646 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12647 throws com.google.protobuf.InvalidProtocolBufferException { 12648 return PARSER.parseFrom(data, extensionRegistry); 12649 } 12650 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(byte[] data) 12651 throws com.google.protobuf.InvalidProtocolBufferException { 12652 return PARSER.parseFrom(data); 12653 } 12654 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12655 byte[] data, 12656 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12657 throws com.google.protobuf.InvalidProtocolBufferException { 12658 return PARSER.parseFrom(data, extensionRegistry); 12659 } 12660 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(java.io.InputStream input) 12661 throws java.io.IOException { 12662 return PARSER.parseFrom(input); 12663 } 12664 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12665 java.io.InputStream input, 12666 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12667 throws java.io.IOException { 12668 return PARSER.parseFrom(input, extensionRegistry); 12669 } 12670 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(java.io.InputStream input) 12671 throws java.io.IOException { 12672 return PARSER.parseDelimitedFrom(input); 12673 } 12674 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom( 12675 java.io.InputStream input, 12676 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12677 throws java.io.IOException { 12678 return PARSER.parseDelimitedFrom(input, extensionRegistry); 12679 } 12680 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12681 com.google.protobuf.CodedInputStream input) 12682 throws java.io.IOException { 12683 return PARSER.parseFrom(input); 12684 } 12685 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom( 12686 com.google.protobuf.CodedInputStream input, 12687 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12688 throws java.io.IOException { 12689 return PARSER.parseFrom(input, extensionRegistry); 12690 } 12691 12692 public static Builder newBuilder() { return Builder.create(); } 12693 public Builder newBuilderForType() { return newBuilder(); } 12694 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto prototype) { 12695 return newBuilder().mergeFrom(prototype); 12696 } 12697 public Builder toBuilder() { return newBuilder(this); } 12698 12699 @java.lang.Override 12700 protected Builder newBuilderForType( 12701 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12702 Builder builder = new Builder(parent); 12703 return builder; 12704 } 12705 /** 12706 * Protobuf type {@code hadoop.hdfs.NewEpochResponseProto} 12707 */ 12708 public static final class Builder extends 12709 com.google.protobuf.GeneratedMessage.Builder<Builder> 12710 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProtoOrBuilder { 12711 public static final com.google.protobuf.Descriptors.Descriptor 12712 getDescriptor() { 12713 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor; 12714 } 12715 12716 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12717 internalGetFieldAccessorTable() { 12718 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable 12719 .ensureFieldAccessorsInitialized( 12720 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class); 12721 } 12722 12723 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.newBuilder() 12724 private Builder() { 12725 maybeForceBuilderInitialization(); 12726 } 12727 12728 private Builder( 12729 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12730 super(parent); 12731 maybeForceBuilderInitialization(); 12732 } 12733 private void maybeForceBuilderInitialization() { 12734 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 12735 } 12736 } 12737 private static Builder create() { 12738 return new Builder(); 12739 } 12740 12741 public Builder clear() { 12742 super.clear(); 12743 lastSegmentTxId_ = 0L; 12744 bitField0_ = (bitField0_ & ~0x00000001); 12745 return this; 12746 } 12747 12748 public Builder clone() { 12749 return create().mergeFrom(buildPartial()); 12750 } 12751 12752 public com.google.protobuf.Descriptors.Descriptor 12753 getDescriptorForType() { 12754 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor; 12755 } 12756 12757 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto getDefaultInstanceForType() { 12758 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(); 12759 } 12760 12761 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto build() { 12762 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = buildPartial(); 12763 if (!result.isInitialized()) { 12764 throw newUninitializedMessageException(result); 12765 } 12766 return result; 12767 } 12768 12769 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto buildPartial() { 12770 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto(this); 12771 int from_bitField0_ = bitField0_; 12772 int to_bitField0_ = 0; 12773 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 12774 to_bitField0_ |= 0x00000001; 12775 } 12776 result.lastSegmentTxId_ = lastSegmentTxId_; 12777 result.bitField0_ = to_bitField0_; 12778 onBuilt(); 12779 return result; 12780 } 12781 12782 public Builder mergeFrom(com.google.protobuf.Message other) { 12783 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) { 12784 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)other); 12785 } else { 12786 super.mergeFrom(other); 12787 return this; 12788 } 12789 } 12790 12791 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other) { 12792 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()) return this; 12793 if (other.hasLastSegmentTxId()) { 12794 setLastSegmentTxId(other.getLastSegmentTxId()); 12795 } 12796 this.mergeUnknownFields(other.getUnknownFields()); 12797 return this; 12798 } 12799 12800 public final boolean isInitialized() { 12801 return true; 12802 } 12803 12804 public Builder mergeFrom( 12805 com.google.protobuf.CodedInputStream input, 12806 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12807 throws java.io.IOException { 12808 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parsedMessage = null; 12809 try { 12810 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 12811 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12812 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) e.getUnfinishedMessage(); 12813 throw e; 12814 } finally { 12815 if (parsedMessage != null) { 12816 mergeFrom(parsedMessage); 12817 } 12818 } 12819 return this; 12820 } 12821 private int bitField0_; 12822 12823 // optional uint64 lastSegmentTxId = 1; 12824 private long lastSegmentTxId_ ; 12825 /** 12826 * <code>optional uint64 lastSegmentTxId = 1;</code> 12827 */ 12828 public boolean hasLastSegmentTxId() { 12829 return ((bitField0_ & 0x00000001) == 0x00000001); 12830 } 12831 /** 12832 * <code>optional uint64 lastSegmentTxId = 1;</code> 12833 */ 12834 public long getLastSegmentTxId() { 12835 return lastSegmentTxId_; 12836 } 12837 /** 12838 * <code>optional uint64 lastSegmentTxId = 1;</code> 12839 */ 12840 public Builder setLastSegmentTxId(long value) { 12841 bitField0_ |= 0x00000001; 12842 lastSegmentTxId_ = value; 12843 onChanged(); 12844 return this; 12845 } 12846 /** 12847 * <code>optional uint64 lastSegmentTxId = 1;</code> 12848 */ 12849 public Builder clearLastSegmentTxId() { 12850 bitField0_ = (bitField0_ & ~0x00000001); 12851 lastSegmentTxId_ = 0L; 12852 onChanged(); 12853 return this; 12854 } 12855 12856 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochResponseProto) 12857 } 12858 12859 static { 12860 defaultInstance = new NewEpochResponseProto(true); 12861 defaultInstance.initFields(); 12862 } 12863 12864 // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochResponseProto) 12865 } 12866 12867 public interface GetEditLogManifestRequestProtoOrBuilder 12868 extends com.google.protobuf.MessageOrBuilder { 12869 12870 // required .hadoop.hdfs.JournalIdProto jid = 1; 12871 /** 12872 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12873 */ 12874 boolean hasJid(); 12875 /** 12876 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12877 */ 12878 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid(); 12879 /** 12880 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 12881 */ 12882 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder(); 12883 12884 // required uint64 sinceTxId = 2; 12885 /** 12886 * <code>required uint64 sinceTxId = 2;</code> 12887 * 12888 * <pre> 12889 * Transaction ID 12890 * </pre> 12891 */ 12892 boolean hasSinceTxId(); 12893 /** 12894 * <code>required uint64 sinceTxId = 2;</code> 12895 * 12896 * <pre> 12897 * Transaction ID 12898 * </pre> 12899 */ 12900 long getSinceTxId(); 12901 12902 // optional bool inProgressOk = 4 [default = false]; 12903 /** 12904 * <code>optional bool inProgressOk = 4 [default = false];</code> 12905 * 12906 * <pre> 12907 * Whether or not the client will be reading from the returned streams. 12908 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 12909 * </pre> 12910 */ 12911 boolean hasInProgressOk(); 12912 /** 12913 * <code>optional bool inProgressOk = 4 [default = false];</code> 12914 * 12915 * <pre> 12916 * Whether or not the client will be reading from the returned streams. 12917 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 12918 * </pre> 12919 */ 12920 boolean getInProgressOk(); 12921 } 12922 /** 12923 * Protobuf type {@code hadoop.hdfs.GetEditLogManifestRequestProto} 12924 * 12925 * <pre> 12926 ** 12927 * getEditLogManifest() 12928 * </pre> 12929 */ 12930 public static final class GetEditLogManifestRequestProto extends 12931 com.google.protobuf.GeneratedMessage 12932 implements GetEditLogManifestRequestProtoOrBuilder { 12933 // Use GetEditLogManifestRequestProto.newBuilder() to construct. 12934 private GetEditLogManifestRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 12935 super(builder); 12936 this.unknownFields = builder.getUnknownFields(); 12937 } 12938 private GetEditLogManifestRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 12939 12940 private static final GetEditLogManifestRequestProto defaultInstance; 12941 public static GetEditLogManifestRequestProto getDefaultInstance() { 12942 return defaultInstance; 12943 } 12944 12945 public GetEditLogManifestRequestProto getDefaultInstanceForType() { 12946 return defaultInstance; 12947 } 12948 12949 private final com.google.protobuf.UnknownFieldSet unknownFields; 12950 @java.lang.Override 12951 public final com.google.protobuf.UnknownFieldSet 12952 getUnknownFields() { 12953 return this.unknownFields; 12954 } 12955 private GetEditLogManifestRequestProto( 12956 com.google.protobuf.CodedInputStream input, 12957 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12958 throws com.google.protobuf.InvalidProtocolBufferException { 12959 initFields(); 12960 int mutable_bitField0_ = 0; 12961 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 12962 com.google.protobuf.UnknownFieldSet.newBuilder(); 12963 try { 12964 boolean done = false; 12965 while (!done) { 12966 int tag = input.readTag(); 12967 switch (tag) { 12968 case 0: 12969 done = true; 12970 break; 12971 default: { 12972 if (!parseUnknownField(input, unknownFields, 12973 extensionRegistry, tag)) { 12974 done = true; 12975 } 12976 break; 12977 } 12978 case 10: { 12979 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null; 12980 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12981 subBuilder = jid_.toBuilder(); 12982 } 12983 jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry); 12984 if (subBuilder != null) { 12985 subBuilder.mergeFrom(jid_); 12986 jid_ = subBuilder.buildPartial(); 12987 } 12988 bitField0_ |= 0x00000001; 12989 break; 12990 } 12991 case 16: { 12992 bitField0_ |= 0x00000002; 12993 sinceTxId_ = input.readUInt64(); 12994 break; 12995 } 12996 case 32: { 12997 bitField0_ |= 0x00000004; 12998 inProgressOk_ = input.readBool(); 12999 break; 13000 } 13001 } 13002 } 13003 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13004 throw e.setUnfinishedMessage(this); 13005 } catch (java.io.IOException e) { 13006 throw new com.google.protobuf.InvalidProtocolBufferException( 13007 e.getMessage()).setUnfinishedMessage(this); 13008 } finally { 13009 this.unknownFields = unknownFields.build(); 13010 makeExtensionsImmutable(); 13011 } 13012 } 13013 public static final com.google.protobuf.Descriptors.Descriptor 13014 getDescriptor() { 13015 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor; 13016 } 13017 13018 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13019 internalGetFieldAccessorTable() { 13020 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable 13021 .ensureFieldAccessorsInitialized( 13022 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class); 13023 } 13024 13025 public static com.google.protobuf.Parser<GetEditLogManifestRequestProto> PARSER = 13026 new com.google.protobuf.AbstractParser<GetEditLogManifestRequestProto>() { 13027 public GetEditLogManifestRequestProto parsePartialFrom( 13028 com.google.protobuf.CodedInputStream input, 13029 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13030 throws com.google.protobuf.InvalidProtocolBufferException { 13031 return new GetEditLogManifestRequestProto(input, extensionRegistry); 13032 } 13033 }; 13034 13035 @java.lang.Override 13036 public com.google.protobuf.Parser<GetEditLogManifestRequestProto> getParserForType() { 13037 return PARSER; 13038 } 13039 13040 private int bitField0_; 13041 // required .hadoop.hdfs.JournalIdProto jid = 1; 13042 public static final int JID_FIELD_NUMBER = 1; 13043 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_; 13044 /** 13045 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13046 */ 13047 public boolean hasJid() { 13048 return ((bitField0_ & 0x00000001) == 0x00000001); 13049 } 13050 /** 13051 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13052 */ 13053 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 13054 return jid_; 13055 } 13056 /** 13057 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13058 */ 13059 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 13060 return jid_; 13061 } 13062 13063 // required uint64 sinceTxId = 2; 13064 public static final int SINCETXID_FIELD_NUMBER = 2; 13065 private long sinceTxId_; 13066 /** 13067 * <code>required uint64 sinceTxId = 2;</code> 13068 * 13069 * <pre> 13070 * Transaction ID 13071 * </pre> 13072 */ 13073 public boolean hasSinceTxId() { 13074 return ((bitField0_ & 0x00000002) == 0x00000002); 13075 } 13076 /** 13077 * <code>required uint64 sinceTxId = 2;</code> 13078 * 13079 * <pre> 13080 * Transaction ID 13081 * </pre> 13082 */ 13083 public long getSinceTxId() { 13084 return sinceTxId_; 13085 } 13086 13087 // optional bool inProgressOk = 4 [default = false]; 13088 public static final int INPROGRESSOK_FIELD_NUMBER = 4; 13089 private boolean inProgressOk_; 13090 /** 13091 * <code>optional bool inProgressOk = 4 [default = false];</code> 13092 * 13093 * <pre> 13094 * Whether or not the client will be reading from the returned streams. 13095 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13096 * </pre> 13097 */ 13098 public boolean hasInProgressOk() { 13099 return ((bitField0_ & 0x00000004) == 0x00000004); 13100 } 13101 /** 13102 * <code>optional bool inProgressOk = 4 [default = false];</code> 13103 * 13104 * <pre> 13105 * Whether or not the client will be reading from the returned streams. 13106 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13107 * </pre> 13108 */ 13109 public boolean getInProgressOk() { 13110 return inProgressOk_; 13111 } 13112 13113 private void initFields() { 13114 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 13115 sinceTxId_ = 0L; 13116 inProgressOk_ = false; 13117 } 13118 private byte memoizedIsInitialized = -1; 13119 public final boolean isInitialized() { 13120 byte isInitialized = memoizedIsInitialized; 13121 if (isInitialized != -1) return isInitialized == 1; 13122 13123 if (!hasJid()) { 13124 memoizedIsInitialized = 0; 13125 return false; 13126 } 13127 if (!hasSinceTxId()) { 13128 memoizedIsInitialized = 0; 13129 return false; 13130 } 13131 if (!getJid().isInitialized()) { 13132 memoizedIsInitialized = 0; 13133 return false; 13134 } 13135 memoizedIsInitialized = 1; 13136 return true; 13137 } 13138 13139 public void writeTo(com.google.protobuf.CodedOutputStream output) 13140 throws java.io.IOException { 13141 getSerializedSize(); 13142 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13143 output.writeMessage(1, jid_); 13144 } 13145 if (((bitField0_ & 0x00000002) == 0x00000002)) { 13146 output.writeUInt64(2, sinceTxId_); 13147 } 13148 if (((bitField0_ & 0x00000004) == 0x00000004)) { 13149 output.writeBool(4, inProgressOk_); 13150 } 13151 getUnknownFields().writeTo(output); 13152 } 13153 13154 private int memoizedSerializedSize = -1; 13155 public int getSerializedSize() { 13156 int size = memoizedSerializedSize; 13157 if (size != -1) return size; 13158 13159 size = 0; 13160 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13161 size += com.google.protobuf.CodedOutputStream 13162 .computeMessageSize(1, jid_); 13163 } 13164 if (((bitField0_ & 0x00000002) == 0x00000002)) { 13165 size += com.google.protobuf.CodedOutputStream 13166 .computeUInt64Size(2, sinceTxId_); 13167 } 13168 if (((bitField0_ & 0x00000004) == 0x00000004)) { 13169 size += com.google.protobuf.CodedOutputStream 13170 .computeBoolSize(4, inProgressOk_); 13171 } 13172 size += getUnknownFields().getSerializedSize(); 13173 memoizedSerializedSize = size; 13174 return size; 13175 } 13176 13177 private static final long serialVersionUID = 0L; 13178 @java.lang.Override 13179 protected java.lang.Object writeReplace() 13180 throws java.io.ObjectStreamException { 13181 return super.writeReplace(); 13182 } 13183 13184 @java.lang.Override 13185 public boolean equals(final java.lang.Object obj) { 13186 if (obj == this) { 13187 return true; 13188 } 13189 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)) { 13190 return super.equals(obj); 13191 } 13192 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) obj; 13193 13194 boolean result = true; 13195 result = result && (hasJid() == other.hasJid()); 13196 if (hasJid()) { 13197 result = result && getJid() 13198 .equals(other.getJid()); 13199 } 13200 result = result && (hasSinceTxId() == other.hasSinceTxId()); 13201 if (hasSinceTxId()) { 13202 result = result && (getSinceTxId() 13203 == other.getSinceTxId()); 13204 } 13205 result = result && (hasInProgressOk() == other.hasInProgressOk()); 13206 if (hasInProgressOk()) { 13207 result = result && (getInProgressOk() 13208 == other.getInProgressOk()); 13209 } 13210 result = result && 13211 getUnknownFields().equals(other.getUnknownFields()); 13212 return result; 13213 } 13214 13215 private int memoizedHashCode = 0; 13216 @java.lang.Override 13217 public int hashCode() { 13218 if (memoizedHashCode != 0) { 13219 return memoizedHashCode; 13220 } 13221 int hash = 41; 13222 hash = (19 * hash) + getDescriptorForType().hashCode(); 13223 if (hasJid()) { 13224 hash = (37 * hash) + JID_FIELD_NUMBER; 13225 hash = (53 * hash) + getJid().hashCode(); 13226 } 13227 if (hasSinceTxId()) { 13228 hash = (37 * hash) + SINCETXID_FIELD_NUMBER; 13229 hash = (53 * hash) + hashLong(getSinceTxId()); 13230 } 13231 if (hasInProgressOk()) { 13232 hash = (37 * hash) + INPROGRESSOK_FIELD_NUMBER; 13233 hash = (53 * hash) + hashBoolean(getInProgressOk()); 13234 } 13235 hash = (29 * hash) + getUnknownFields().hashCode(); 13236 memoizedHashCode = hash; 13237 return hash; 13238 } 13239 13240 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13241 com.google.protobuf.ByteString data) 13242 throws com.google.protobuf.InvalidProtocolBufferException { 13243 return PARSER.parseFrom(data); 13244 } 13245 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13246 com.google.protobuf.ByteString data, 13247 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13248 throws com.google.protobuf.InvalidProtocolBufferException { 13249 return PARSER.parseFrom(data, extensionRegistry); 13250 } 13251 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(byte[] data) 13252 throws com.google.protobuf.InvalidProtocolBufferException { 13253 return PARSER.parseFrom(data); 13254 } 13255 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13256 byte[] data, 13257 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13258 throws com.google.protobuf.InvalidProtocolBufferException { 13259 return PARSER.parseFrom(data, extensionRegistry); 13260 } 13261 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(java.io.InputStream input) 13262 throws java.io.IOException { 13263 return PARSER.parseFrom(input); 13264 } 13265 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13266 java.io.InputStream input, 13267 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13268 throws java.io.IOException { 13269 return PARSER.parseFrom(input, extensionRegistry); 13270 } 13271 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(java.io.InputStream input) 13272 throws java.io.IOException { 13273 return PARSER.parseDelimitedFrom(input); 13274 } 13275 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom( 13276 java.io.InputStream input, 13277 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13278 throws java.io.IOException { 13279 return PARSER.parseDelimitedFrom(input, extensionRegistry); 13280 } 13281 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13282 com.google.protobuf.CodedInputStream input) 13283 throws java.io.IOException { 13284 return PARSER.parseFrom(input); 13285 } 13286 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom( 13287 com.google.protobuf.CodedInputStream input, 13288 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13289 throws java.io.IOException { 13290 return PARSER.parseFrom(input, extensionRegistry); 13291 } 13292 13293 public static Builder newBuilder() { return Builder.create(); } 13294 public Builder newBuilderForType() { return newBuilder(); } 13295 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto prototype) { 13296 return newBuilder().mergeFrom(prototype); 13297 } 13298 public Builder toBuilder() { return newBuilder(this); } 13299 13300 @java.lang.Override 13301 protected Builder newBuilderForType( 13302 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13303 Builder builder = new Builder(parent); 13304 return builder; 13305 } 13306 /** 13307 * Protobuf type {@code hadoop.hdfs.GetEditLogManifestRequestProto} 13308 * 13309 * <pre> 13310 ** 13311 * getEditLogManifest() 13312 * </pre> 13313 */ 13314 public static final class Builder extends 13315 com.google.protobuf.GeneratedMessage.Builder<Builder> 13316 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProtoOrBuilder { 13317 public static final com.google.protobuf.Descriptors.Descriptor 13318 getDescriptor() { 13319 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor; 13320 } 13321 13322 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13323 internalGetFieldAccessorTable() { 13324 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable 13325 .ensureFieldAccessorsInitialized( 13326 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class); 13327 } 13328 13329 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.newBuilder() 13330 private Builder() { 13331 maybeForceBuilderInitialization(); 13332 } 13333 13334 private Builder( 13335 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13336 super(parent); 13337 maybeForceBuilderInitialization(); 13338 } 13339 private void maybeForceBuilderInitialization() { 13340 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 13341 getJidFieldBuilder(); 13342 } 13343 } 13344 private static Builder create() { 13345 return new Builder(); 13346 } 13347 13348 public Builder clear() { 13349 super.clear(); 13350 if (jidBuilder_ == null) { 13351 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 13352 } else { 13353 jidBuilder_.clear(); 13354 } 13355 bitField0_ = (bitField0_ & ~0x00000001); 13356 sinceTxId_ = 0L; 13357 bitField0_ = (bitField0_ & ~0x00000002); 13358 inProgressOk_ = false; 13359 bitField0_ = (bitField0_ & ~0x00000004); 13360 return this; 13361 } 13362 13363 public Builder clone() { 13364 return create().mergeFrom(buildPartial()); 13365 } 13366 13367 public com.google.protobuf.Descriptors.Descriptor 13368 getDescriptorForType() { 13369 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor; 13370 } 13371 13372 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto getDefaultInstanceForType() { 13373 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance(); 13374 } 13375 13376 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto build() { 13377 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = buildPartial(); 13378 if (!result.isInitialized()) { 13379 throw newUninitializedMessageException(result); 13380 } 13381 return result; 13382 } 13383 13384 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto buildPartial() { 13385 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto(this); 13386 int from_bitField0_ = bitField0_; 13387 int to_bitField0_ = 0; 13388 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 13389 to_bitField0_ |= 0x00000001; 13390 } 13391 if (jidBuilder_ == null) { 13392 result.jid_ = jid_; 13393 } else { 13394 result.jid_ = jidBuilder_.build(); 13395 } 13396 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 13397 to_bitField0_ |= 0x00000002; 13398 } 13399 result.sinceTxId_ = sinceTxId_; 13400 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 13401 to_bitField0_ |= 0x00000004; 13402 } 13403 result.inProgressOk_ = inProgressOk_; 13404 result.bitField0_ = to_bitField0_; 13405 onBuilt(); 13406 return result; 13407 } 13408 13409 public Builder mergeFrom(com.google.protobuf.Message other) { 13410 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) { 13411 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)other); 13412 } else { 13413 super.mergeFrom(other); 13414 return this; 13415 } 13416 } 13417 13418 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other) { 13419 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance()) return this; 13420 if (other.hasJid()) { 13421 mergeJid(other.getJid()); 13422 } 13423 if (other.hasSinceTxId()) { 13424 setSinceTxId(other.getSinceTxId()); 13425 } 13426 if (other.hasInProgressOk()) { 13427 setInProgressOk(other.getInProgressOk()); 13428 } 13429 this.mergeUnknownFields(other.getUnknownFields()); 13430 return this; 13431 } 13432 13433 public final boolean isInitialized() { 13434 if (!hasJid()) { 13435 13436 return false; 13437 } 13438 if (!hasSinceTxId()) { 13439 13440 return false; 13441 } 13442 if (!getJid().isInitialized()) { 13443 13444 return false; 13445 } 13446 return true; 13447 } 13448 13449 public Builder mergeFrom( 13450 com.google.protobuf.CodedInputStream input, 13451 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13452 throws java.io.IOException { 13453 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parsedMessage = null; 13454 try { 13455 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 13456 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13457 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) e.getUnfinishedMessage(); 13458 throw e; 13459 } finally { 13460 if (parsedMessage != null) { 13461 mergeFrom(parsedMessage); 13462 } 13463 } 13464 return this; 13465 } 13466 private int bitField0_; 13467 13468 // required .hadoop.hdfs.JournalIdProto jid = 1; 13469 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 13470 private com.google.protobuf.SingleFieldBuilder< 13471 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_; 13472 /** 13473 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13474 */ 13475 public boolean hasJid() { 13476 return ((bitField0_ & 0x00000001) == 0x00000001); 13477 } 13478 /** 13479 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13480 */ 13481 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() { 13482 if (jidBuilder_ == null) { 13483 return jid_; 13484 } else { 13485 return jidBuilder_.getMessage(); 13486 } 13487 } 13488 /** 13489 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13490 */ 13491 public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 13492 if (jidBuilder_ == null) { 13493 if (value == null) { 13494 throw new NullPointerException(); 13495 } 13496 jid_ = value; 13497 onChanged(); 13498 } else { 13499 jidBuilder_.setMessage(value); 13500 } 13501 bitField0_ |= 0x00000001; 13502 return this; 13503 } 13504 /** 13505 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13506 */ 13507 public Builder setJid( 13508 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) { 13509 if (jidBuilder_ == null) { 13510 jid_ = builderForValue.build(); 13511 onChanged(); 13512 } else { 13513 jidBuilder_.setMessage(builderForValue.build()); 13514 } 13515 bitField0_ |= 0x00000001; 13516 return this; 13517 } 13518 /** 13519 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13520 */ 13521 public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) { 13522 if (jidBuilder_ == null) { 13523 if (((bitField0_ & 0x00000001) == 0x00000001) && 13524 jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) { 13525 jid_ = 13526 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial(); 13527 } else { 13528 jid_ = value; 13529 } 13530 onChanged(); 13531 } else { 13532 jidBuilder_.mergeFrom(value); 13533 } 13534 bitField0_ |= 0x00000001; 13535 return this; 13536 } 13537 /** 13538 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13539 */ 13540 public Builder clearJid() { 13541 if (jidBuilder_ == null) { 13542 jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance(); 13543 onChanged(); 13544 } else { 13545 jidBuilder_.clear(); 13546 } 13547 bitField0_ = (bitField0_ & ~0x00000001); 13548 return this; 13549 } 13550 /** 13551 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13552 */ 13553 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() { 13554 bitField0_ |= 0x00000001; 13555 onChanged(); 13556 return getJidFieldBuilder().getBuilder(); 13557 } 13558 /** 13559 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13560 */ 13561 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() { 13562 if (jidBuilder_ != null) { 13563 return jidBuilder_.getMessageOrBuilder(); 13564 } else { 13565 return jid_; 13566 } 13567 } 13568 /** 13569 * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code> 13570 */ 13571 private com.google.protobuf.SingleFieldBuilder< 13572 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 13573 getJidFieldBuilder() { 13574 if (jidBuilder_ == null) { 13575 jidBuilder_ = new com.google.protobuf.SingleFieldBuilder< 13576 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>( 13577 jid_, 13578 getParentForChildren(), 13579 isClean()); 13580 jid_ = null; 13581 } 13582 return jidBuilder_; 13583 } 13584 13585 // required uint64 sinceTxId = 2; 13586 private long sinceTxId_ ; 13587 /** 13588 * <code>required uint64 sinceTxId = 2;</code> 13589 * 13590 * <pre> 13591 * Transaction ID 13592 * </pre> 13593 */ 13594 public boolean hasSinceTxId() { 13595 return ((bitField0_ & 0x00000002) == 0x00000002); 13596 } 13597 /** 13598 * <code>required uint64 sinceTxId = 2;</code> 13599 * 13600 * <pre> 13601 * Transaction ID 13602 * </pre> 13603 */ 13604 public long getSinceTxId() { 13605 return sinceTxId_; 13606 } 13607 /** 13608 * <code>required uint64 sinceTxId = 2;</code> 13609 * 13610 * <pre> 13611 * Transaction ID 13612 * </pre> 13613 */ 13614 public Builder setSinceTxId(long value) { 13615 bitField0_ |= 0x00000002; 13616 sinceTxId_ = value; 13617 onChanged(); 13618 return this; 13619 } 13620 /** 13621 * <code>required uint64 sinceTxId = 2;</code> 13622 * 13623 * <pre> 13624 * Transaction ID 13625 * </pre> 13626 */ 13627 public Builder clearSinceTxId() { 13628 bitField0_ = (bitField0_ & ~0x00000002); 13629 sinceTxId_ = 0L; 13630 onChanged(); 13631 return this; 13632 } 13633 13634 // optional bool inProgressOk = 4 [default = false]; 13635 private boolean inProgressOk_ ; 13636 /** 13637 * <code>optional bool inProgressOk = 4 [default = false];</code> 13638 * 13639 * <pre> 13640 * Whether or not the client will be reading from the returned streams. 13641 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13642 * </pre> 13643 */ 13644 public boolean hasInProgressOk() { 13645 return ((bitField0_ & 0x00000004) == 0x00000004); 13646 } 13647 /** 13648 * <code>optional bool inProgressOk = 4 [default = false];</code> 13649 * 13650 * <pre> 13651 * Whether or not the client will be reading from the returned streams. 13652 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13653 * </pre> 13654 */ 13655 public boolean getInProgressOk() { 13656 return inProgressOk_; 13657 } 13658 /** 13659 * <code>optional bool inProgressOk = 4 [default = false];</code> 13660 * 13661 * <pre> 13662 * Whether or not the client will be reading from the returned streams. 13663 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13664 * </pre> 13665 */ 13666 public Builder setInProgressOk(boolean value) { 13667 bitField0_ |= 0x00000004; 13668 inProgressOk_ = value; 13669 onChanged(); 13670 return this; 13671 } 13672 /** 13673 * <code>optional bool inProgressOk = 4 [default = false];</code> 13674 * 13675 * <pre> 13676 * Whether or not the client will be reading from the returned streams. 13677 * optional bool forReading = 3 [default = true]; <obsolete, do not reuse> 13678 * </pre> 13679 */ 13680 public Builder clearInProgressOk() { 13681 bitField0_ = (bitField0_ & ~0x00000004); 13682 inProgressOk_ = false; 13683 onChanged(); 13684 return this; 13685 } 13686 13687 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestRequestProto) 13688 } 13689 13690 static { 13691 defaultInstance = new GetEditLogManifestRequestProto(true); 13692 defaultInstance.initFields(); 13693 } 13694 13695 // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestRequestProto) 13696 } 13697 13698 public interface GetEditLogManifestResponseProtoOrBuilder 13699 extends com.google.protobuf.MessageOrBuilder { 13700 13701 // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1; 13702 /** 13703 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13704 */ 13705 boolean hasManifest(); 13706 /** 13707 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13708 */ 13709 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest(); 13710 /** 13711 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13712 */ 13713 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder(); 13714 13715 // required uint32 httpPort = 2; 13716 /** 13717 * <code>required uint32 httpPort = 2;</code> 13718 * 13719 * <pre> 13720 * Deprecated by fromURL 13721 * </pre> 13722 */ 13723 boolean hasHttpPort(); 13724 /** 13725 * <code>required uint32 httpPort = 2;</code> 13726 * 13727 * <pre> 13728 * Deprecated by fromURL 13729 * </pre> 13730 */ 13731 int getHttpPort(); 13732 13733 // optional string fromURL = 3; 13734 /** 13735 * <code>optional string fromURL = 3;</code> 13736 */ 13737 boolean hasFromURL(); 13738 /** 13739 * <code>optional string fromURL = 3;</code> 13740 */ 13741 java.lang.String getFromURL(); 13742 /** 13743 * <code>optional string fromURL = 3;</code> 13744 */ 13745 com.google.protobuf.ByteString 13746 getFromURLBytes(); 13747 } 13748 /** 13749 * Protobuf type {@code hadoop.hdfs.GetEditLogManifestResponseProto} 13750 */ 13751 public static final class GetEditLogManifestResponseProto extends 13752 com.google.protobuf.GeneratedMessage 13753 implements GetEditLogManifestResponseProtoOrBuilder { 13754 // Use GetEditLogManifestResponseProto.newBuilder() to construct. 13755 private GetEditLogManifestResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13756 super(builder); 13757 this.unknownFields = builder.getUnknownFields(); 13758 } 13759 private GetEditLogManifestResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13760 13761 private static final GetEditLogManifestResponseProto defaultInstance; 13762 public static GetEditLogManifestResponseProto getDefaultInstance() { 13763 return defaultInstance; 13764 } 13765 13766 public GetEditLogManifestResponseProto getDefaultInstanceForType() { 13767 return defaultInstance; 13768 } 13769 13770 private final com.google.protobuf.UnknownFieldSet unknownFields; 13771 @java.lang.Override 13772 public final com.google.protobuf.UnknownFieldSet 13773 getUnknownFields() { 13774 return this.unknownFields; 13775 } 13776 private GetEditLogManifestResponseProto( 13777 com.google.protobuf.CodedInputStream input, 13778 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13779 throws com.google.protobuf.InvalidProtocolBufferException { 13780 initFields(); 13781 int mutable_bitField0_ = 0; 13782 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13783 com.google.protobuf.UnknownFieldSet.newBuilder(); 13784 try { 13785 boolean done = false; 13786 while (!done) { 13787 int tag = input.readTag(); 13788 switch (tag) { 13789 case 0: 13790 done = true; 13791 break; 13792 default: { 13793 if (!parseUnknownField(input, unknownFields, 13794 extensionRegistry, tag)) { 13795 done = true; 13796 } 13797 break; 13798 } 13799 case 10: { 13800 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder subBuilder = null; 13801 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13802 subBuilder = manifest_.toBuilder(); 13803 } 13804 manifest_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.PARSER, extensionRegistry); 13805 if (subBuilder != null) { 13806 subBuilder.mergeFrom(manifest_); 13807 manifest_ = subBuilder.buildPartial(); 13808 } 13809 bitField0_ |= 0x00000001; 13810 break; 13811 } 13812 case 16: { 13813 bitField0_ |= 0x00000002; 13814 httpPort_ = input.readUInt32(); 13815 break; 13816 } 13817 case 26: { 13818 bitField0_ |= 0x00000004; 13819 fromURL_ = input.readBytes(); 13820 break; 13821 } 13822 } 13823 } 13824 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13825 throw e.setUnfinishedMessage(this); 13826 } catch (java.io.IOException e) { 13827 throw new com.google.protobuf.InvalidProtocolBufferException( 13828 e.getMessage()).setUnfinishedMessage(this); 13829 } finally { 13830 this.unknownFields = unknownFields.build(); 13831 makeExtensionsImmutable(); 13832 } 13833 } 13834 public static final com.google.protobuf.Descriptors.Descriptor 13835 getDescriptor() { 13836 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor; 13837 } 13838 13839 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13840 internalGetFieldAccessorTable() { 13841 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable 13842 .ensureFieldAccessorsInitialized( 13843 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class); 13844 } 13845 13846 public static com.google.protobuf.Parser<GetEditLogManifestResponseProto> PARSER = 13847 new com.google.protobuf.AbstractParser<GetEditLogManifestResponseProto>() { 13848 public GetEditLogManifestResponseProto parsePartialFrom( 13849 com.google.protobuf.CodedInputStream input, 13850 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13851 throws com.google.protobuf.InvalidProtocolBufferException { 13852 return new GetEditLogManifestResponseProto(input, extensionRegistry); 13853 } 13854 }; 13855 13856 @java.lang.Override 13857 public com.google.protobuf.Parser<GetEditLogManifestResponseProto> getParserForType() { 13858 return PARSER; 13859 } 13860 13861 private int bitField0_; 13862 // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1; 13863 public static final int MANIFEST_FIELD_NUMBER = 1; 13864 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_; 13865 /** 13866 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13867 */ 13868 public boolean hasManifest() { 13869 return ((bitField0_ & 0x00000001) == 0x00000001); 13870 } 13871 /** 13872 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13873 */ 13874 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() { 13875 return manifest_; 13876 } 13877 /** 13878 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 13879 */ 13880 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() { 13881 return manifest_; 13882 } 13883 13884 // required uint32 httpPort = 2; 13885 public static final int HTTPPORT_FIELD_NUMBER = 2; 13886 private int httpPort_; 13887 /** 13888 * <code>required uint32 httpPort = 2;</code> 13889 * 13890 * <pre> 13891 * Deprecated by fromURL 13892 * </pre> 13893 */ 13894 public boolean hasHttpPort() { 13895 return ((bitField0_ & 0x00000002) == 0x00000002); 13896 } 13897 /** 13898 * <code>required uint32 httpPort = 2;</code> 13899 * 13900 * <pre> 13901 * Deprecated by fromURL 13902 * </pre> 13903 */ 13904 public int getHttpPort() { 13905 return httpPort_; 13906 } 13907 13908 // optional string fromURL = 3; 13909 public static final int FROMURL_FIELD_NUMBER = 3; 13910 private java.lang.Object fromURL_; 13911 /** 13912 * <code>optional string fromURL = 3;</code> 13913 */ 13914 public boolean hasFromURL() { 13915 return ((bitField0_ & 0x00000004) == 0x00000004); 13916 } 13917 /** 13918 * <code>optional string fromURL = 3;</code> 13919 */ 13920 public java.lang.String getFromURL() { 13921 java.lang.Object ref = fromURL_; 13922 if (ref instanceof java.lang.String) { 13923 return (java.lang.String) ref; 13924 } else { 13925 com.google.protobuf.ByteString bs = 13926 (com.google.protobuf.ByteString) ref; 13927 java.lang.String s = bs.toStringUtf8(); 13928 if (bs.isValidUtf8()) { 13929 fromURL_ = s; 13930 } 13931 return s; 13932 } 13933 } 13934 /** 13935 * <code>optional string fromURL = 3;</code> 13936 */ 13937 public com.google.protobuf.ByteString 13938 getFromURLBytes() { 13939 java.lang.Object ref = fromURL_; 13940 if (ref instanceof java.lang.String) { 13941 com.google.protobuf.ByteString b = 13942 com.google.protobuf.ByteString.copyFromUtf8( 13943 (java.lang.String) ref); 13944 fromURL_ = b; 13945 return b; 13946 } else { 13947 return (com.google.protobuf.ByteString) ref; 13948 } 13949 } 13950 13951 private void initFields() { 13952 manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance(); 13953 httpPort_ = 0; 13954 fromURL_ = ""; 13955 } 13956 private byte memoizedIsInitialized = -1; 13957 public final boolean isInitialized() { 13958 byte isInitialized = memoizedIsInitialized; 13959 if (isInitialized != -1) return isInitialized == 1; 13960 13961 if (!hasManifest()) { 13962 memoizedIsInitialized = 0; 13963 return false; 13964 } 13965 if (!hasHttpPort()) { 13966 memoizedIsInitialized = 0; 13967 return false; 13968 } 13969 if (!getManifest().isInitialized()) { 13970 memoizedIsInitialized = 0; 13971 return false; 13972 } 13973 memoizedIsInitialized = 1; 13974 return true; 13975 } 13976 13977 public void writeTo(com.google.protobuf.CodedOutputStream output) 13978 throws java.io.IOException { 13979 getSerializedSize(); 13980 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13981 output.writeMessage(1, manifest_); 13982 } 13983 if (((bitField0_ & 0x00000002) == 0x00000002)) { 13984 output.writeUInt32(2, httpPort_); 13985 } 13986 if (((bitField0_ & 0x00000004) == 0x00000004)) { 13987 output.writeBytes(3, getFromURLBytes()); 13988 } 13989 getUnknownFields().writeTo(output); 13990 } 13991 13992 private int memoizedSerializedSize = -1; 13993 public int getSerializedSize() { 13994 int size = memoizedSerializedSize; 13995 if (size != -1) return size; 13996 13997 size = 0; 13998 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13999 size += com.google.protobuf.CodedOutputStream 14000 .computeMessageSize(1, manifest_); 14001 } 14002 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14003 size += com.google.protobuf.CodedOutputStream 14004 .computeUInt32Size(2, httpPort_); 14005 } 14006 if (((bitField0_ & 0x00000004) == 0x00000004)) { 14007 size += com.google.protobuf.CodedOutputStream 14008 .computeBytesSize(3, getFromURLBytes()); 14009 } 14010 size += getUnknownFields().getSerializedSize(); 14011 memoizedSerializedSize = size; 14012 return size; 14013 } 14014 14015 private static final long serialVersionUID = 0L; 14016 @java.lang.Override 14017 protected java.lang.Object writeReplace() 14018 throws java.io.ObjectStreamException { 14019 return super.writeReplace(); 14020 } 14021 14022 @java.lang.Override 14023 public boolean equals(final java.lang.Object obj) { 14024 if (obj == this) { 14025 return true; 14026 } 14027 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)) { 14028 return super.equals(obj); 14029 } 14030 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) obj; 14031 14032 boolean result = true; 14033 result = result && (hasManifest() == other.hasManifest()); 14034 if (hasManifest()) { 14035 result = result && getManifest() 14036 .equals(other.getManifest()); 14037 } 14038 result = result && (hasHttpPort() == other.hasHttpPort()); 14039 if (hasHttpPort()) { 14040 result = result && (getHttpPort() 14041 == other.getHttpPort()); 14042 } 14043 result = result && (hasFromURL() == other.hasFromURL()); 14044 if (hasFromURL()) { 14045 result = result && getFromURL() 14046 .equals(other.getFromURL()); 14047 } 14048 result = result && 14049 getUnknownFields().equals(other.getUnknownFields()); 14050 return result; 14051 } 14052 14053 private int memoizedHashCode = 0; 14054 @java.lang.Override 14055 public int hashCode() { 14056 if (memoizedHashCode != 0) { 14057 return memoizedHashCode; 14058 } 14059 int hash = 41; 14060 hash = (19 * hash) + getDescriptorForType().hashCode(); 14061 if (hasManifest()) { 14062 hash = (37 * hash) + MANIFEST_FIELD_NUMBER; 14063 hash = (53 * hash) + getManifest().hashCode(); 14064 } 14065 if (hasHttpPort()) { 14066 hash = (37 * hash) + HTTPPORT_FIELD_NUMBER; 14067 hash = (53 * hash) + getHttpPort(); 14068 } 14069 if (hasFromURL()) { 14070 hash = (37 * hash) + FROMURL_FIELD_NUMBER; 14071 hash = (53 * hash) + getFromURL().hashCode(); 14072 } 14073 hash = (29 * hash) + getUnknownFields().hashCode(); 14074 memoizedHashCode = hash; 14075 return hash; 14076 } 14077 14078 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14079 com.google.protobuf.ByteString data) 14080 throws com.google.protobuf.InvalidProtocolBufferException { 14081 return PARSER.parseFrom(data); 14082 } 14083 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14084 com.google.protobuf.ByteString data, 14085 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14086 throws com.google.protobuf.InvalidProtocolBufferException { 14087 return PARSER.parseFrom(data, extensionRegistry); 14088 } 14089 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(byte[] data) 14090 throws com.google.protobuf.InvalidProtocolBufferException { 14091 return PARSER.parseFrom(data); 14092 } 14093 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14094 byte[] data, 14095 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14096 throws com.google.protobuf.InvalidProtocolBufferException { 14097 return PARSER.parseFrom(data, extensionRegistry); 14098 } 14099 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(java.io.InputStream input) 14100 throws java.io.IOException { 14101 return PARSER.parseFrom(input); 14102 } 14103 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14104 java.io.InputStream input, 14105 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14106 throws java.io.IOException { 14107 return PARSER.parseFrom(input, extensionRegistry); 14108 } 14109 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(java.io.InputStream input) 14110 throws java.io.IOException { 14111 return PARSER.parseDelimitedFrom(input); 14112 } 14113 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom( 14114 java.io.InputStream input, 14115 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14116 throws java.io.IOException { 14117 return PARSER.parseDelimitedFrom(input, extensionRegistry); 14118 } 14119 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14120 com.google.protobuf.CodedInputStream input) 14121 throws java.io.IOException { 14122 return PARSER.parseFrom(input); 14123 } 14124 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom( 14125 com.google.protobuf.CodedInputStream input, 14126 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14127 throws java.io.IOException { 14128 return PARSER.parseFrom(input, extensionRegistry); 14129 } 14130 14131 public static Builder newBuilder() { return Builder.create(); } 14132 public Builder newBuilderForType() { return newBuilder(); } 14133 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto prototype) { 14134 return newBuilder().mergeFrom(prototype); 14135 } 14136 public Builder toBuilder() { return newBuilder(this); } 14137 14138 @java.lang.Override 14139 protected Builder newBuilderForType( 14140 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14141 Builder builder = new Builder(parent); 14142 return builder; 14143 } 14144 /** 14145 * Protobuf type {@code hadoop.hdfs.GetEditLogManifestResponseProto} 14146 */ 14147 public static final class Builder extends 14148 com.google.protobuf.GeneratedMessage.Builder<Builder> 14149 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProtoOrBuilder { 14150 public static final com.google.protobuf.Descriptors.Descriptor 14151 getDescriptor() { 14152 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor; 14153 } 14154 14155 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14156 internalGetFieldAccessorTable() { 14157 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable 14158 .ensureFieldAccessorsInitialized( 14159 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class); 14160 } 14161 14162 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.newBuilder() 14163 private Builder() { 14164 maybeForceBuilderInitialization(); 14165 } 14166 14167 private Builder( 14168 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14169 super(parent); 14170 maybeForceBuilderInitialization(); 14171 } 14172 private void maybeForceBuilderInitialization() { 14173 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 14174 getManifestFieldBuilder(); 14175 } 14176 } 14177 private static Builder create() { 14178 return new Builder(); 14179 } 14180 14181 public Builder clear() { 14182 super.clear(); 14183 if (manifestBuilder_ == null) { 14184 manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance(); 14185 } else { 14186 manifestBuilder_.clear(); 14187 } 14188 bitField0_ = (bitField0_ & ~0x00000001); 14189 httpPort_ = 0; 14190 bitField0_ = (bitField0_ & ~0x00000002); 14191 fromURL_ = ""; 14192 bitField0_ = (bitField0_ & ~0x00000004); 14193 return this; 14194 } 14195 14196 public Builder clone() { 14197 return create().mergeFrom(buildPartial()); 14198 } 14199 14200 public com.google.protobuf.Descriptors.Descriptor 14201 getDescriptorForType() { 14202 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor; 14203 } 14204 14205 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getDefaultInstanceForType() { 14206 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(); 14207 } 14208 14209 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto build() { 14210 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = buildPartial(); 14211 if (!result.isInitialized()) { 14212 throw newUninitializedMessageException(result); 14213 } 14214 return result; 14215 } 14216 14217 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto buildPartial() { 14218 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto(this); 14219 int from_bitField0_ = bitField0_; 14220 int to_bitField0_ = 0; 14221 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 14222 to_bitField0_ |= 0x00000001; 14223 } 14224 if (manifestBuilder_ == null) { 14225 result.manifest_ = manifest_; 14226 } else { 14227 result.manifest_ = manifestBuilder_.build(); 14228 } 14229 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 14230 to_bitField0_ |= 0x00000002; 14231 } 14232 result.httpPort_ = httpPort_; 14233 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 14234 to_bitField0_ |= 0x00000004; 14235 } 14236 result.fromURL_ = fromURL_; 14237 result.bitField0_ = to_bitField0_; 14238 onBuilt(); 14239 return result; 14240 } 14241 14242 public Builder mergeFrom(com.google.protobuf.Message other) { 14243 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) { 14244 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)other); 14245 } else { 14246 super.mergeFrom(other); 14247 return this; 14248 } 14249 } 14250 14251 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other) { 14252 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()) return this; 14253 if (other.hasManifest()) { 14254 mergeManifest(other.getManifest()); 14255 } 14256 if (other.hasHttpPort()) { 14257 setHttpPort(other.getHttpPort()); 14258 } 14259 if (other.hasFromURL()) { 14260 bitField0_ |= 0x00000004; 14261 fromURL_ = other.fromURL_; 14262 onChanged(); 14263 } 14264 this.mergeUnknownFields(other.getUnknownFields()); 14265 return this; 14266 } 14267 14268 public final boolean isInitialized() { 14269 if (!hasManifest()) { 14270 14271 return false; 14272 } 14273 if (!hasHttpPort()) { 14274 14275 return false; 14276 } 14277 if (!getManifest().isInitialized()) { 14278 14279 return false; 14280 } 14281 return true; 14282 } 14283 14284 public Builder mergeFrom( 14285 com.google.protobuf.CodedInputStream input, 14286 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14287 throws java.io.IOException { 14288 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parsedMessage = null; 14289 try { 14290 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 14291 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 14292 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) e.getUnfinishedMessage(); 14293 throw e; 14294 } finally { 14295 if (parsedMessage != null) { 14296 mergeFrom(parsedMessage); 14297 } 14298 } 14299 return this; 14300 } 14301 private int bitField0_; 14302 14303 // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1; 14304 private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance(); 14305 private com.google.protobuf.SingleFieldBuilder< 14306 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> manifestBuilder_; 14307 /** 14308 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14309 */ 14310 public boolean hasManifest() { 14311 return ((bitField0_ & 0x00000001) == 0x00000001); 14312 } 14313 /** 14314 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14315 */ 14316 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() { 14317 if (manifestBuilder_ == null) { 14318 return manifest_; 14319 } else { 14320 return manifestBuilder_.getMessage(); 14321 } 14322 } 14323 /** 14324 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14325 */ 14326 public Builder setManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) { 14327 if (manifestBuilder_ == null) { 14328 if (value == null) { 14329 throw new NullPointerException(); 14330 } 14331 manifest_ = value; 14332 onChanged(); 14333 } else { 14334 manifestBuilder_.setMessage(value); 14335 } 14336 bitField0_ |= 0x00000001; 14337 return this; 14338 } 14339 /** 14340 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14341 */ 14342 public Builder setManifest( 14343 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder builderForValue) { 14344 if (manifestBuilder_ == null) { 14345 manifest_ = builderForValue.build(); 14346 onChanged(); 14347 } else { 14348 manifestBuilder_.setMessage(builderForValue.build()); 14349 } 14350 bitField0_ |= 0x00000001; 14351 return this; 14352 } 14353 /** 14354 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14355 */ 14356 public Builder mergeManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) { 14357 if (manifestBuilder_ == null) { 14358 if (((bitField0_ & 0x00000001) == 0x00000001) && 14359 manifest_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance()) { 14360 manifest_ = 14361 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.newBuilder(manifest_).mergeFrom(value).buildPartial(); 14362 } else { 14363 manifest_ = value; 14364 } 14365 onChanged(); 14366 } else { 14367 manifestBuilder_.mergeFrom(value); 14368 } 14369 bitField0_ |= 0x00000001; 14370 return this; 14371 } 14372 /** 14373 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14374 */ 14375 public Builder clearManifest() { 14376 if (manifestBuilder_ == null) { 14377 manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance(); 14378 onChanged(); 14379 } else { 14380 manifestBuilder_.clear(); 14381 } 14382 bitField0_ = (bitField0_ & ~0x00000001); 14383 return this; 14384 } 14385 /** 14386 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14387 */ 14388 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder getManifestBuilder() { 14389 bitField0_ |= 0x00000001; 14390 onChanged(); 14391 return getManifestFieldBuilder().getBuilder(); 14392 } 14393 /** 14394 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14395 */ 14396 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() { 14397 if (manifestBuilder_ != null) { 14398 return manifestBuilder_.getMessageOrBuilder(); 14399 } else { 14400 return manifest_; 14401 } 14402 } 14403 /** 14404 * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code> 14405 */ 14406 private com.google.protobuf.SingleFieldBuilder< 14407 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> 14408 getManifestFieldBuilder() { 14409 if (manifestBuilder_ == null) { 14410 manifestBuilder_ = new com.google.protobuf.SingleFieldBuilder< 14411 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder>( 14412 manifest_, 14413 getParentForChildren(), 14414 isClean()); 14415 manifest_ = null; 14416 } 14417 return manifestBuilder_; 14418 } 14419 14420 // required uint32 httpPort = 2; 14421 private int httpPort_ ; 14422 /** 14423 * <code>required uint32 httpPort = 2;</code> 14424 * 14425 * <pre> 14426 * Deprecated by fromURL 14427 * </pre> 14428 */ 14429 public boolean hasHttpPort() { 14430 return ((bitField0_ & 0x00000002) == 0x00000002); 14431 } 14432 /** 14433 * <code>required uint32 httpPort = 2;</code> 14434 * 14435 * <pre> 14436 * Deprecated by fromURL 14437 * </pre> 14438 */ 14439 public int getHttpPort() { 14440 return httpPort_; 14441 } 14442 /** 14443 * <code>required uint32 httpPort = 2;</code> 14444 * 14445 * <pre> 14446 * Deprecated by fromURL 14447 * </pre> 14448 */ 14449 public Builder setHttpPort(int value) { 14450 bitField0_ |= 0x00000002; 14451 httpPort_ = value; 14452 onChanged(); 14453 return this; 14454 } 14455 /** 14456 * <code>required uint32 httpPort = 2;</code> 14457 * 14458 * <pre> 14459 * Deprecated by fromURL 14460 * </pre> 14461 */ 14462 public Builder clearHttpPort() { 14463 bitField0_ = (bitField0_ & ~0x00000002); 14464 httpPort_ = 0; 14465 onChanged(); 14466 return this; 14467 } 14468 14469 // optional string fromURL = 3; 14470 private java.lang.Object fromURL_ = ""; 14471 /** 14472 * <code>optional string fromURL = 3;</code> 14473 */ 14474 public boolean hasFromURL() { 14475 return ((bitField0_ & 0x00000004) == 0x00000004); 14476 } 14477 /** 14478 * <code>optional string fromURL = 3;</code> 14479 */ 14480 public java.lang.String getFromURL() { 14481 java.lang.Object ref = fromURL_; 14482 if (!(ref instanceof java.lang.String)) { 14483 java.lang.String s = ((com.google.protobuf.ByteString) ref) 14484 .toStringUtf8(); 14485 fromURL_ = s; 14486 return s; 14487 } else { 14488 return (java.lang.String) ref; 14489 } 14490 } 14491 /** 14492 * <code>optional string fromURL = 3;</code> 14493 */ 14494 public com.google.protobuf.ByteString 14495 getFromURLBytes() { 14496 java.lang.Object ref = fromURL_; 14497 if (ref instanceof String) { 14498 com.google.protobuf.ByteString b = 14499 com.google.protobuf.ByteString.copyFromUtf8( 14500 (java.lang.String) ref); 14501 fromURL_ = b; 14502 return b; 14503 } else { 14504 return (com.google.protobuf.ByteString) ref; 14505 } 14506 } 14507 /** 14508 * <code>optional string fromURL = 3;</code> 14509 */ 14510 public Builder setFromURL( 14511 java.lang.String value) { 14512 if (value == null) { 14513 throw new NullPointerException(); 14514 } 14515 bitField0_ |= 0x00000004; 14516 fromURL_ = value; 14517 onChanged(); 14518 return this; 14519 } 14520 /** 14521 * <code>optional string fromURL = 3;</code> 14522 */ 14523 public Builder clearFromURL() { 14524 bitField0_ = (bitField0_ & ~0x00000004); 14525 fromURL_ = getDefaultInstance().getFromURL(); 14526 onChanged(); 14527 return this; 14528 } 14529 /** 14530 * <code>optional string fromURL = 3;</code> 14531 */ 14532 public Builder setFromURLBytes( 14533 com.google.protobuf.ByteString value) { 14534 if (value == null) { 14535 throw new NullPointerException(); 14536 } 14537 bitField0_ |= 0x00000004; 14538 fromURL_ = value; 14539 onChanged(); 14540 return this; 14541 } 14542 14543 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestResponseProto) 14544 } 14545 14546 static { 14547 defaultInstance = new GetEditLogManifestResponseProto(true); 14548 defaultInstance.initFields(); 14549 } 14550 14551 // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestResponseProto) 14552 } 14553 14554 public interface PrepareRecoveryRequestProtoOrBuilder 14555 extends com.google.protobuf.MessageOrBuilder { 14556 14557 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 14558 /** 14559 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14560 */ 14561 boolean hasReqInfo(); 14562 /** 14563 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14564 */ 14565 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 14566 /** 14567 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14568 */ 14569 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 14570 14571 // required uint64 segmentTxId = 2; 14572 /** 14573 * <code>required uint64 segmentTxId = 2;</code> 14574 */ 14575 boolean hasSegmentTxId(); 14576 /** 14577 * <code>required uint64 segmentTxId = 2;</code> 14578 */ 14579 long getSegmentTxId(); 14580 } 14581 /** 14582 * Protobuf type {@code hadoop.hdfs.PrepareRecoveryRequestProto} 14583 * 14584 * <pre> 14585 ** 14586 * prepareRecovery() 14587 * </pre> 14588 */ 14589 public static final class PrepareRecoveryRequestProto extends 14590 com.google.protobuf.GeneratedMessage 14591 implements PrepareRecoveryRequestProtoOrBuilder { 14592 // Use PrepareRecoveryRequestProto.newBuilder() to construct. 14593 private PrepareRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 14594 super(builder); 14595 this.unknownFields = builder.getUnknownFields(); 14596 } 14597 private PrepareRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 14598 14599 private static final PrepareRecoveryRequestProto defaultInstance; 14600 public static PrepareRecoveryRequestProto getDefaultInstance() { 14601 return defaultInstance; 14602 } 14603 14604 public PrepareRecoveryRequestProto getDefaultInstanceForType() { 14605 return defaultInstance; 14606 } 14607 14608 private final com.google.protobuf.UnknownFieldSet unknownFields; 14609 @java.lang.Override 14610 public final com.google.protobuf.UnknownFieldSet 14611 getUnknownFields() { 14612 return this.unknownFields; 14613 } 14614 private PrepareRecoveryRequestProto( 14615 com.google.protobuf.CodedInputStream input, 14616 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14617 throws com.google.protobuf.InvalidProtocolBufferException { 14618 initFields(); 14619 int mutable_bitField0_ = 0; 14620 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 14621 com.google.protobuf.UnknownFieldSet.newBuilder(); 14622 try { 14623 boolean done = false; 14624 while (!done) { 14625 int tag = input.readTag(); 14626 switch (tag) { 14627 case 0: 14628 done = true; 14629 break; 14630 default: { 14631 if (!parseUnknownField(input, unknownFields, 14632 extensionRegistry, tag)) { 14633 done = true; 14634 } 14635 break; 14636 } 14637 case 10: { 14638 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 14639 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14640 subBuilder = reqInfo_.toBuilder(); 14641 } 14642 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 14643 if (subBuilder != null) { 14644 subBuilder.mergeFrom(reqInfo_); 14645 reqInfo_ = subBuilder.buildPartial(); 14646 } 14647 bitField0_ |= 0x00000001; 14648 break; 14649 } 14650 case 16: { 14651 bitField0_ |= 0x00000002; 14652 segmentTxId_ = input.readUInt64(); 14653 break; 14654 } 14655 } 14656 } 14657 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 14658 throw e.setUnfinishedMessage(this); 14659 } catch (java.io.IOException e) { 14660 throw new com.google.protobuf.InvalidProtocolBufferException( 14661 e.getMessage()).setUnfinishedMessage(this); 14662 } finally { 14663 this.unknownFields = unknownFields.build(); 14664 makeExtensionsImmutable(); 14665 } 14666 } 14667 public static final com.google.protobuf.Descriptors.Descriptor 14668 getDescriptor() { 14669 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor; 14670 } 14671 14672 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14673 internalGetFieldAccessorTable() { 14674 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable 14675 .ensureFieldAccessorsInitialized( 14676 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class); 14677 } 14678 14679 public static com.google.protobuf.Parser<PrepareRecoveryRequestProto> PARSER = 14680 new com.google.protobuf.AbstractParser<PrepareRecoveryRequestProto>() { 14681 public PrepareRecoveryRequestProto parsePartialFrom( 14682 com.google.protobuf.CodedInputStream input, 14683 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14684 throws com.google.protobuf.InvalidProtocolBufferException { 14685 return new PrepareRecoveryRequestProto(input, extensionRegistry); 14686 } 14687 }; 14688 14689 @java.lang.Override 14690 public com.google.protobuf.Parser<PrepareRecoveryRequestProto> getParserForType() { 14691 return PARSER; 14692 } 14693 14694 private int bitField0_; 14695 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 14696 public static final int REQINFO_FIELD_NUMBER = 1; 14697 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 14698 /** 14699 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14700 */ 14701 public boolean hasReqInfo() { 14702 return ((bitField0_ & 0x00000001) == 0x00000001); 14703 } 14704 /** 14705 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14706 */ 14707 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 14708 return reqInfo_; 14709 } 14710 /** 14711 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 14712 */ 14713 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 14714 return reqInfo_; 14715 } 14716 14717 // required uint64 segmentTxId = 2; 14718 public static final int SEGMENTTXID_FIELD_NUMBER = 2; 14719 private long segmentTxId_; 14720 /** 14721 * <code>required uint64 segmentTxId = 2;</code> 14722 */ 14723 public boolean hasSegmentTxId() { 14724 return ((bitField0_ & 0x00000002) == 0x00000002); 14725 } 14726 /** 14727 * <code>required uint64 segmentTxId = 2;</code> 14728 */ 14729 public long getSegmentTxId() { 14730 return segmentTxId_; 14731 } 14732 14733 private void initFields() { 14734 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 14735 segmentTxId_ = 0L; 14736 } 14737 private byte memoizedIsInitialized = -1; 14738 public final boolean isInitialized() { 14739 byte isInitialized = memoizedIsInitialized; 14740 if (isInitialized != -1) return isInitialized == 1; 14741 14742 if (!hasReqInfo()) { 14743 memoizedIsInitialized = 0; 14744 return false; 14745 } 14746 if (!hasSegmentTxId()) { 14747 memoizedIsInitialized = 0; 14748 return false; 14749 } 14750 if (!getReqInfo().isInitialized()) { 14751 memoizedIsInitialized = 0; 14752 return false; 14753 } 14754 memoizedIsInitialized = 1; 14755 return true; 14756 } 14757 14758 public void writeTo(com.google.protobuf.CodedOutputStream output) 14759 throws java.io.IOException { 14760 getSerializedSize(); 14761 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14762 output.writeMessage(1, reqInfo_); 14763 } 14764 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14765 output.writeUInt64(2, segmentTxId_); 14766 } 14767 getUnknownFields().writeTo(output); 14768 } 14769 14770 private int memoizedSerializedSize = -1; 14771 public int getSerializedSize() { 14772 int size = memoizedSerializedSize; 14773 if (size != -1) return size; 14774 14775 size = 0; 14776 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14777 size += com.google.protobuf.CodedOutputStream 14778 .computeMessageSize(1, reqInfo_); 14779 } 14780 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14781 size += com.google.protobuf.CodedOutputStream 14782 .computeUInt64Size(2, segmentTxId_); 14783 } 14784 size += getUnknownFields().getSerializedSize(); 14785 memoizedSerializedSize = size; 14786 return size; 14787 } 14788 14789 private static final long serialVersionUID = 0L; 14790 @java.lang.Override 14791 protected java.lang.Object writeReplace() 14792 throws java.io.ObjectStreamException { 14793 return super.writeReplace(); 14794 } 14795 14796 @java.lang.Override 14797 public boolean equals(final java.lang.Object obj) { 14798 if (obj == this) { 14799 return true; 14800 } 14801 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)) { 14802 return super.equals(obj); 14803 } 14804 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) obj; 14805 14806 boolean result = true; 14807 result = result && (hasReqInfo() == other.hasReqInfo()); 14808 if (hasReqInfo()) { 14809 result = result && getReqInfo() 14810 .equals(other.getReqInfo()); 14811 } 14812 result = result && (hasSegmentTxId() == other.hasSegmentTxId()); 14813 if (hasSegmentTxId()) { 14814 result = result && (getSegmentTxId() 14815 == other.getSegmentTxId()); 14816 } 14817 result = result && 14818 getUnknownFields().equals(other.getUnknownFields()); 14819 return result; 14820 } 14821 14822 private int memoizedHashCode = 0; 14823 @java.lang.Override 14824 public int hashCode() { 14825 if (memoizedHashCode != 0) { 14826 return memoizedHashCode; 14827 } 14828 int hash = 41; 14829 hash = (19 * hash) + getDescriptorForType().hashCode(); 14830 if (hasReqInfo()) { 14831 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 14832 hash = (53 * hash) + getReqInfo().hashCode(); 14833 } 14834 if (hasSegmentTxId()) { 14835 hash = (37 * hash) + SEGMENTTXID_FIELD_NUMBER; 14836 hash = (53 * hash) + hashLong(getSegmentTxId()); 14837 } 14838 hash = (29 * hash) + getUnknownFields().hashCode(); 14839 memoizedHashCode = hash; 14840 return hash; 14841 } 14842 14843 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14844 com.google.protobuf.ByteString data) 14845 throws com.google.protobuf.InvalidProtocolBufferException { 14846 return PARSER.parseFrom(data); 14847 } 14848 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14849 com.google.protobuf.ByteString data, 14850 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14851 throws com.google.protobuf.InvalidProtocolBufferException { 14852 return PARSER.parseFrom(data, extensionRegistry); 14853 } 14854 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(byte[] data) 14855 throws com.google.protobuf.InvalidProtocolBufferException { 14856 return PARSER.parseFrom(data); 14857 } 14858 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14859 byte[] data, 14860 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14861 throws com.google.protobuf.InvalidProtocolBufferException { 14862 return PARSER.parseFrom(data, extensionRegistry); 14863 } 14864 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(java.io.InputStream input) 14865 throws java.io.IOException { 14866 return PARSER.parseFrom(input); 14867 } 14868 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14869 java.io.InputStream input, 14870 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14871 throws java.io.IOException { 14872 return PARSER.parseFrom(input, extensionRegistry); 14873 } 14874 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input) 14875 throws java.io.IOException { 14876 return PARSER.parseDelimitedFrom(input); 14877 } 14878 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom( 14879 java.io.InputStream input, 14880 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14881 throws java.io.IOException { 14882 return PARSER.parseDelimitedFrom(input, extensionRegistry); 14883 } 14884 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14885 com.google.protobuf.CodedInputStream input) 14886 throws java.io.IOException { 14887 return PARSER.parseFrom(input); 14888 } 14889 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom( 14890 com.google.protobuf.CodedInputStream input, 14891 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14892 throws java.io.IOException { 14893 return PARSER.parseFrom(input, extensionRegistry); 14894 } 14895 14896 public static Builder newBuilder() { return Builder.create(); } 14897 public Builder newBuilderForType() { return newBuilder(); } 14898 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto prototype) { 14899 return newBuilder().mergeFrom(prototype); 14900 } 14901 public Builder toBuilder() { return newBuilder(this); } 14902 14903 @java.lang.Override 14904 protected Builder newBuilderForType( 14905 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14906 Builder builder = new Builder(parent); 14907 return builder; 14908 } 14909 /** 14910 * Protobuf type {@code hadoop.hdfs.PrepareRecoveryRequestProto} 14911 * 14912 * <pre> 14913 ** 14914 * prepareRecovery() 14915 * </pre> 14916 */ 14917 public static final class Builder extends 14918 com.google.protobuf.GeneratedMessage.Builder<Builder> 14919 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProtoOrBuilder { 14920 public static final com.google.protobuf.Descriptors.Descriptor 14921 getDescriptor() { 14922 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor; 14923 } 14924 14925 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14926 internalGetFieldAccessorTable() { 14927 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable 14928 .ensureFieldAccessorsInitialized( 14929 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class); 14930 } 14931 14932 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.newBuilder() 14933 private Builder() { 14934 maybeForceBuilderInitialization(); 14935 } 14936 14937 private Builder( 14938 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14939 super(parent); 14940 maybeForceBuilderInitialization(); 14941 } 14942 private void maybeForceBuilderInitialization() { 14943 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 14944 getReqInfoFieldBuilder(); 14945 } 14946 } 14947 private static Builder create() { 14948 return new Builder(); 14949 } 14950 14951 public Builder clear() { 14952 super.clear(); 14953 if (reqInfoBuilder_ == null) { 14954 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 14955 } else { 14956 reqInfoBuilder_.clear(); 14957 } 14958 bitField0_ = (bitField0_ & ~0x00000001); 14959 segmentTxId_ = 0L; 14960 bitField0_ = (bitField0_ & ~0x00000002); 14961 return this; 14962 } 14963 14964 public Builder clone() { 14965 return create().mergeFrom(buildPartial()); 14966 } 14967 14968 public com.google.protobuf.Descriptors.Descriptor 14969 getDescriptorForType() { 14970 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor; 14971 } 14972 14973 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto getDefaultInstanceForType() { 14974 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance(); 14975 } 14976 14977 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto build() { 14978 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = buildPartial(); 14979 if (!result.isInitialized()) { 14980 throw newUninitializedMessageException(result); 14981 } 14982 return result; 14983 } 14984 14985 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto buildPartial() { 14986 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto(this); 14987 int from_bitField0_ = bitField0_; 14988 int to_bitField0_ = 0; 14989 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 14990 to_bitField0_ |= 0x00000001; 14991 } 14992 if (reqInfoBuilder_ == null) { 14993 result.reqInfo_ = reqInfo_; 14994 } else { 14995 result.reqInfo_ = reqInfoBuilder_.build(); 14996 } 14997 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 14998 to_bitField0_ |= 0x00000002; 14999 } 15000 result.segmentTxId_ = segmentTxId_; 15001 result.bitField0_ = to_bitField0_; 15002 onBuilt(); 15003 return result; 15004 } 15005 15006 public Builder mergeFrom(com.google.protobuf.Message other) { 15007 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) { 15008 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)other); 15009 } else { 15010 super.mergeFrom(other); 15011 return this; 15012 } 15013 } 15014 15015 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other) { 15016 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance()) return this; 15017 if (other.hasReqInfo()) { 15018 mergeReqInfo(other.getReqInfo()); 15019 } 15020 if (other.hasSegmentTxId()) { 15021 setSegmentTxId(other.getSegmentTxId()); 15022 } 15023 this.mergeUnknownFields(other.getUnknownFields()); 15024 return this; 15025 } 15026 15027 public final boolean isInitialized() { 15028 if (!hasReqInfo()) { 15029 15030 return false; 15031 } 15032 if (!hasSegmentTxId()) { 15033 15034 return false; 15035 } 15036 if (!getReqInfo().isInitialized()) { 15037 15038 return false; 15039 } 15040 return true; 15041 } 15042 15043 public Builder mergeFrom( 15044 com.google.protobuf.CodedInputStream input, 15045 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15046 throws java.io.IOException { 15047 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parsedMessage = null; 15048 try { 15049 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 15050 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15051 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) e.getUnfinishedMessage(); 15052 throw e; 15053 } finally { 15054 if (parsedMessage != null) { 15055 mergeFrom(parsedMessage); 15056 } 15057 } 15058 return this; 15059 } 15060 private int bitField0_; 15061 15062 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 15063 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 15064 private com.google.protobuf.SingleFieldBuilder< 15065 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 15066 /** 15067 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15068 */ 15069 public boolean hasReqInfo() { 15070 return ((bitField0_ & 0x00000001) == 0x00000001); 15071 } 15072 /** 15073 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15074 */ 15075 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 15076 if (reqInfoBuilder_ == null) { 15077 return reqInfo_; 15078 } else { 15079 return reqInfoBuilder_.getMessage(); 15080 } 15081 } 15082 /** 15083 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15084 */ 15085 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 15086 if (reqInfoBuilder_ == null) { 15087 if (value == null) { 15088 throw new NullPointerException(); 15089 } 15090 reqInfo_ = value; 15091 onChanged(); 15092 } else { 15093 reqInfoBuilder_.setMessage(value); 15094 } 15095 bitField0_ |= 0x00000001; 15096 return this; 15097 } 15098 /** 15099 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15100 */ 15101 public Builder setReqInfo( 15102 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 15103 if (reqInfoBuilder_ == null) { 15104 reqInfo_ = builderForValue.build(); 15105 onChanged(); 15106 } else { 15107 reqInfoBuilder_.setMessage(builderForValue.build()); 15108 } 15109 bitField0_ |= 0x00000001; 15110 return this; 15111 } 15112 /** 15113 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15114 */ 15115 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 15116 if (reqInfoBuilder_ == null) { 15117 if (((bitField0_ & 0x00000001) == 0x00000001) && 15118 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 15119 reqInfo_ = 15120 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 15121 } else { 15122 reqInfo_ = value; 15123 } 15124 onChanged(); 15125 } else { 15126 reqInfoBuilder_.mergeFrom(value); 15127 } 15128 bitField0_ |= 0x00000001; 15129 return this; 15130 } 15131 /** 15132 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15133 */ 15134 public Builder clearReqInfo() { 15135 if (reqInfoBuilder_ == null) { 15136 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 15137 onChanged(); 15138 } else { 15139 reqInfoBuilder_.clear(); 15140 } 15141 bitField0_ = (bitField0_ & ~0x00000001); 15142 return this; 15143 } 15144 /** 15145 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15146 */ 15147 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 15148 bitField0_ |= 0x00000001; 15149 onChanged(); 15150 return getReqInfoFieldBuilder().getBuilder(); 15151 } 15152 /** 15153 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15154 */ 15155 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 15156 if (reqInfoBuilder_ != null) { 15157 return reqInfoBuilder_.getMessageOrBuilder(); 15158 } else { 15159 return reqInfo_; 15160 } 15161 } 15162 /** 15163 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 15164 */ 15165 private com.google.protobuf.SingleFieldBuilder< 15166 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 15167 getReqInfoFieldBuilder() { 15168 if (reqInfoBuilder_ == null) { 15169 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 15170 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 15171 reqInfo_, 15172 getParentForChildren(), 15173 isClean()); 15174 reqInfo_ = null; 15175 } 15176 return reqInfoBuilder_; 15177 } 15178 15179 // required uint64 segmentTxId = 2; 15180 private long segmentTxId_ ; 15181 /** 15182 * <code>required uint64 segmentTxId = 2;</code> 15183 */ 15184 public boolean hasSegmentTxId() { 15185 return ((bitField0_ & 0x00000002) == 0x00000002); 15186 } 15187 /** 15188 * <code>required uint64 segmentTxId = 2;</code> 15189 */ 15190 public long getSegmentTxId() { 15191 return segmentTxId_; 15192 } 15193 /** 15194 * <code>required uint64 segmentTxId = 2;</code> 15195 */ 15196 public Builder setSegmentTxId(long value) { 15197 bitField0_ |= 0x00000002; 15198 segmentTxId_ = value; 15199 onChanged(); 15200 return this; 15201 } 15202 /** 15203 * <code>required uint64 segmentTxId = 2;</code> 15204 */ 15205 public Builder clearSegmentTxId() { 15206 bitField0_ = (bitField0_ & ~0x00000002); 15207 segmentTxId_ = 0L; 15208 onChanged(); 15209 return this; 15210 } 15211 15212 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryRequestProto) 15213 } 15214 15215 static { 15216 defaultInstance = new PrepareRecoveryRequestProto(true); 15217 defaultInstance.initFields(); 15218 } 15219 15220 // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryRequestProto) 15221 } 15222 15223 public interface PrepareRecoveryResponseProtoOrBuilder 15224 extends com.google.protobuf.MessageOrBuilder { 15225 15226 // optional .hadoop.hdfs.SegmentStateProto segmentState = 1; 15227 /** 15228 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15229 */ 15230 boolean hasSegmentState(); 15231 /** 15232 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15233 */ 15234 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState(); 15235 /** 15236 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15237 */ 15238 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder(); 15239 15240 // optional uint64 acceptedInEpoch = 2; 15241 /** 15242 * <code>optional uint64 acceptedInEpoch = 2;</code> 15243 */ 15244 boolean hasAcceptedInEpoch(); 15245 /** 15246 * <code>optional uint64 acceptedInEpoch = 2;</code> 15247 */ 15248 long getAcceptedInEpoch(); 15249 15250 // required uint64 lastWriterEpoch = 3; 15251 /** 15252 * <code>required uint64 lastWriterEpoch = 3;</code> 15253 */ 15254 boolean hasLastWriterEpoch(); 15255 /** 15256 * <code>required uint64 lastWriterEpoch = 3;</code> 15257 */ 15258 long getLastWriterEpoch(); 15259 15260 // optional uint64 lastCommittedTxId = 4; 15261 /** 15262 * <code>optional uint64 lastCommittedTxId = 4;</code> 15263 * 15264 * <pre> 15265 * The highest committed txid that this logger has ever seen. 15266 * This may be higher than the data it actually has, in the case 15267 * that it was lagging before the old writer crashed. 15268 * </pre> 15269 */ 15270 boolean hasLastCommittedTxId(); 15271 /** 15272 * <code>optional uint64 lastCommittedTxId = 4;</code> 15273 * 15274 * <pre> 15275 * The highest committed txid that this logger has ever seen. 15276 * This may be higher than the data it actually has, in the case 15277 * that it was lagging before the old writer crashed. 15278 * </pre> 15279 */ 15280 long getLastCommittedTxId(); 15281 } 15282 /** 15283 * Protobuf type {@code hadoop.hdfs.PrepareRecoveryResponseProto} 15284 */ 15285 public static final class PrepareRecoveryResponseProto extends 15286 com.google.protobuf.GeneratedMessage 15287 implements PrepareRecoveryResponseProtoOrBuilder { 15288 // Use PrepareRecoveryResponseProto.newBuilder() to construct. 15289 private PrepareRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 15290 super(builder); 15291 this.unknownFields = builder.getUnknownFields(); 15292 } 15293 private PrepareRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 15294 15295 private static final PrepareRecoveryResponseProto defaultInstance; 15296 public static PrepareRecoveryResponseProto getDefaultInstance() { 15297 return defaultInstance; 15298 } 15299 15300 public PrepareRecoveryResponseProto getDefaultInstanceForType() { 15301 return defaultInstance; 15302 } 15303 15304 private final com.google.protobuf.UnknownFieldSet unknownFields; 15305 @java.lang.Override 15306 public final com.google.protobuf.UnknownFieldSet 15307 getUnknownFields() { 15308 return this.unknownFields; 15309 } 15310 private PrepareRecoveryResponseProto( 15311 com.google.protobuf.CodedInputStream input, 15312 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15313 throws com.google.protobuf.InvalidProtocolBufferException { 15314 initFields(); 15315 int mutable_bitField0_ = 0; 15316 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15317 com.google.protobuf.UnknownFieldSet.newBuilder(); 15318 try { 15319 boolean done = false; 15320 while (!done) { 15321 int tag = input.readTag(); 15322 switch (tag) { 15323 case 0: 15324 done = true; 15325 break; 15326 default: { 15327 if (!parseUnknownField(input, unknownFields, 15328 extensionRegistry, tag)) { 15329 done = true; 15330 } 15331 break; 15332 } 15333 case 10: { 15334 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null; 15335 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15336 subBuilder = segmentState_.toBuilder(); 15337 } 15338 segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry); 15339 if (subBuilder != null) { 15340 subBuilder.mergeFrom(segmentState_); 15341 segmentState_ = subBuilder.buildPartial(); 15342 } 15343 bitField0_ |= 0x00000001; 15344 break; 15345 } 15346 case 16: { 15347 bitField0_ |= 0x00000002; 15348 acceptedInEpoch_ = input.readUInt64(); 15349 break; 15350 } 15351 case 24: { 15352 bitField0_ |= 0x00000004; 15353 lastWriterEpoch_ = input.readUInt64(); 15354 break; 15355 } 15356 case 32: { 15357 bitField0_ |= 0x00000008; 15358 lastCommittedTxId_ = input.readUInt64(); 15359 break; 15360 } 15361 } 15362 } 15363 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15364 throw e.setUnfinishedMessage(this); 15365 } catch (java.io.IOException e) { 15366 throw new com.google.protobuf.InvalidProtocolBufferException( 15367 e.getMessage()).setUnfinishedMessage(this); 15368 } finally { 15369 this.unknownFields = unknownFields.build(); 15370 makeExtensionsImmutable(); 15371 } 15372 } 15373 public static final com.google.protobuf.Descriptors.Descriptor 15374 getDescriptor() { 15375 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor; 15376 } 15377 15378 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15379 internalGetFieldAccessorTable() { 15380 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable 15381 .ensureFieldAccessorsInitialized( 15382 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class); 15383 } 15384 15385 public static com.google.protobuf.Parser<PrepareRecoveryResponseProto> PARSER = 15386 new com.google.protobuf.AbstractParser<PrepareRecoveryResponseProto>() { 15387 public PrepareRecoveryResponseProto parsePartialFrom( 15388 com.google.protobuf.CodedInputStream input, 15389 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15390 throws com.google.protobuf.InvalidProtocolBufferException { 15391 return new PrepareRecoveryResponseProto(input, extensionRegistry); 15392 } 15393 }; 15394 15395 @java.lang.Override 15396 public com.google.protobuf.Parser<PrepareRecoveryResponseProto> getParserForType() { 15397 return PARSER; 15398 } 15399 15400 private int bitField0_; 15401 // optional .hadoop.hdfs.SegmentStateProto segmentState = 1; 15402 public static final int SEGMENTSTATE_FIELD_NUMBER = 1; 15403 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_; 15404 /** 15405 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15406 */ 15407 public boolean hasSegmentState() { 15408 return ((bitField0_ & 0x00000001) == 0x00000001); 15409 } 15410 /** 15411 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15412 */ 15413 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() { 15414 return segmentState_; 15415 } 15416 /** 15417 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15418 */ 15419 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() { 15420 return segmentState_; 15421 } 15422 15423 // optional uint64 acceptedInEpoch = 2; 15424 public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2; 15425 private long acceptedInEpoch_; 15426 /** 15427 * <code>optional uint64 acceptedInEpoch = 2;</code> 15428 */ 15429 public boolean hasAcceptedInEpoch() { 15430 return ((bitField0_ & 0x00000002) == 0x00000002); 15431 } 15432 /** 15433 * <code>optional uint64 acceptedInEpoch = 2;</code> 15434 */ 15435 public long getAcceptedInEpoch() { 15436 return acceptedInEpoch_; 15437 } 15438 15439 // required uint64 lastWriterEpoch = 3; 15440 public static final int LASTWRITEREPOCH_FIELD_NUMBER = 3; 15441 private long lastWriterEpoch_; 15442 /** 15443 * <code>required uint64 lastWriterEpoch = 3;</code> 15444 */ 15445 public boolean hasLastWriterEpoch() { 15446 return ((bitField0_ & 0x00000004) == 0x00000004); 15447 } 15448 /** 15449 * <code>required uint64 lastWriterEpoch = 3;</code> 15450 */ 15451 public long getLastWriterEpoch() { 15452 return lastWriterEpoch_; 15453 } 15454 15455 // optional uint64 lastCommittedTxId = 4; 15456 public static final int LASTCOMMITTEDTXID_FIELD_NUMBER = 4; 15457 private long lastCommittedTxId_; 15458 /** 15459 * <code>optional uint64 lastCommittedTxId = 4;</code> 15460 * 15461 * <pre> 15462 * The highest committed txid that this logger has ever seen. 15463 * This may be higher than the data it actually has, in the case 15464 * that it was lagging before the old writer crashed. 15465 * </pre> 15466 */ 15467 public boolean hasLastCommittedTxId() { 15468 return ((bitField0_ & 0x00000008) == 0x00000008); 15469 } 15470 /** 15471 * <code>optional uint64 lastCommittedTxId = 4;</code> 15472 * 15473 * <pre> 15474 * The highest committed txid that this logger has ever seen. 15475 * This may be higher than the data it actually has, in the case 15476 * that it was lagging before the old writer crashed. 15477 * </pre> 15478 */ 15479 public long getLastCommittedTxId() { 15480 return lastCommittedTxId_; 15481 } 15482 15483 private void initFields() { 15484 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 15485 acceptedInEpoch_ = 0L; 15486 lastWriterEpoch_ = 0L; 15487 lastCommittedTxId_ = 0L; 15488 } 15489 private byte memoizedIsInitialized = -1; 15490 public final boolean isInitialized() { 15491 byte isInitialized = memoizedIsInitialized; 15492 if (isInitialized != -1) return isInitialized == 1; 15493 15494 if (!hasLastWriterEpoch()) { 15495 memoizedIsInitialized = 0; 15496 return false; 15497 } 15498 if (hasSegmentState()) { 15499 if (!getSegmentState().isInitialized()) { 15500 memoizedIsInitialized = 0; 15501 return false; 15502 } 15503 } 15504 memoizedIsInitialized = 1; 15505 return true; 15506 } 15507 15508 public void writeTo(com.google.protobuf.CodedOutputStream output) 15509 throws java.io.IOException { 15510 getSerializedSize(); 15511 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15512 output.writeMessage(1, segmentState_); 15513 } 15514 if (((bitField0_ & 0x00000002) == 0x00000002)) { 15515 output.writeUInt64(2, acceptedInEpoch_); 15516 } 15517 if (((bitField0_ & 0x00000004) == 0x00000004)) { 15518 output.writeUInt64(3, lastWriterEpoch_); 15519 } 15520 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15521 output.writeUInt64(4, lastCommittedTxId_); 15522 } 15523 getUnknownFields().writeTo(output); 15524 } 15525 15526 private int memoizedSerializedSize = -1; 15527 public int getSerializedSize() { 15528 int size = memoizedSerializedSize; 15529 if (size != -1) return size; 15530 15531 size = 0; 15532 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15533 size += com.google.protobuf.CodedOutputStream 15534 .computeMessageSize(1, segmentState_); 15535 } 15536 if (((bitField0_ & 0x00000002) == 0x00000002)) { 15537 size += com.google.protobuf.CodedOutputStream 15538 .computeUInt64Size(2, acceptedInEpoch_); 15539 } 15540 if (((bitField0_ & 0x00000004) == 0x00000004)) { 15541 size += com.google.protobuf.CodedOutputStream 15542 .computeUInt64Size(3, lastWriterEpoch_); 15543 } 15544 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15545 size += com.google.protobuf.CodedOutputStream 15546 .computeUInt64Size(4, lastCommittedTxId_); 15547 } 15548 size += getUnknownFields().getSerializedSize(); 15549 memoizedSerializedSize = size; 15550 return size; 15551 } 15552 15553 private static final long serialVersionUID = 0L; 15554 @java.lang.Override 15555 protected java.lang.Object writeReplace() 15556 throws java.io.ObjectStreamException { 15557 return super.writeReplace(); 15558 } 15559 15560 @java.lang.Override 15561 public boolean equals(final java.lang.Object obj) { 15562 if (obj == this) { 15563 return true; 15564 } 15565 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)) { 15566 return super.equals(obj); 15567 } 15568 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) obj; 15569 15570 boolean result = true; 15571 result = result && (hasSegmentState() == other.hasSegmentState()); 15572 if (hasSegmentState()) { 15573 result = result && getSegmentState() 15574 .equals(other.getSegmentState()); 15575 } 15576 result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch()); 15577 if (hasAcceptedInEpoch()) { 15578 result = result && (getAcceptedInEpoch() 15579 == other.getAcceptedInEpoch()); 15580 } 15581 result = result && (hasLastWriterEpoch() == other.hasLastWriterEpoch()); 15582 if (hasLastWriterEpoch()) { 15583 result = result && (getLastWriterEpoch() 15584 == other.getLastWriterEpoch()); 15585 } 15586 result = result && (hasLastCommittedTxId() == other.hasLastCommittedTxId()); 15587 if (hasLastCommittedTxId()) { 15588 result = result && (getLastCommittedTxId() 15589 == other.getLastCommittedTxId()); 15590 } 15591 result = result && 15592 getUnknownFields().equals(other.getUnknownFields()); 15593 return result; 15594 } 15595 15596 private int memoizedHashCode = 0; 15597 @java.lang.Override 15598 public int hashCode() { 15599 if (memoizedHashCode != 0) { 15600 return memoizedHashCode; 15601 } 15602 int hash = 41; 15603 hash = (19 * hash) + getDescriptorForType().hashCode(); 15604 if (hasSegmentState()) { 15605 hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER; 15606 hash = (53 * hash) + getSegmentState().hashCode(); 15607 } 15608 if (hasAcceptedInEpoch()) { 15609 hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER; 15610 hash = (53 * hash) + hashLong(getAcceptedInEpoch()); 15611 } 15612 if (hasLastWriterEpoch()) { 15613 hash = (37 * hash) + LASTWRITEREPOCH_FIELD_NUMBER; 15614 hash = (53 * hash) + hashLong(getLastWriterEpoch()); 15615 } 15616 if (hasLastCommittedTxId()) { 15617 hash = (37 * hash) + LASTCOMMITTEDTXID_FIELD_NUMBER; 15618 hash = (53 * hash) + hashLong(getLastCommittedTxId()); 15619 } 15620 hash = (29 * hash) + getUnknownFields().hashCode(); 15621 memoizedHashCode = hash; 15622 return hash; 15623 } 15624 15625 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15626 com.google.protobuf.ByteString data) 15627 throws com.google.protobuf.InvalidProtocolBufferException { 15628 return PARSER.parseFrom(data); 15629 } 15630 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15631 com.google.protobuf.ByteString data, 15632 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15633 throws com.google.protobuf.InvalidProtocolBufferException { 15634 return PARSER.parseFrom(data, extensionRegistry); 15635 } 15636 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(byte[] data) 15637 throws com.google.protobuf.InvalidProtocolBufferException { 15638 return PARSER.parseFrom(data); 15639 } 15640 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15641 byte[] data, 15642 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15643 throws com.google.protobuf.InvalidProtocolBufferException { 15644 return PARSER.parseFrom(data, extensionRegistry); 15645 } 15646 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(java.io.InputStream input) 15647 throws java.io.IOException { 15648 return PARSER.parseFrom(input); 15649 } 15650 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15651 java.io.InputStream input, 15652 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15653 throws java.io.IOException { 15654 return PARSER.parseFrom(input, extensionRegistry); 15655 } 15656 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input) 15657 throws java.io.IOException { 15658 return PARSER.parseDelimitedFrom(input); 15659 } 15660 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom( 15661 java.io.InputStream input, 15662 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15663 throws java.io.IOException { 15664 return PARSER.parseDelimitedFrom(input, extensionRegistry); 15665 } 15666 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15667 com.google.protobuf.CodedInputStream input) 15668 throws java.io.IOException { 15669 return PARSER.parseFrom(input); 15670 } 15671 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom( 15672 com.google.protobuf.CodedInputStream input, 15673 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15674 throws java.io.IOException { 15675 return PARSER.parseFrom(input, extensionRegistry); 15676 } 15677 15678 public static Builder newBuilder() { return Builder.create(); } 15679 public Builder newBuilderForType() { return newBuilder(); } 15680 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prototype) { 15681 return newBuilder().mergeFrom(prototype); 15682 } 15683 public Builder toBuilder() { return newBuilder(this); } 15684 15685 @java.lang.Override 15686 protected Builder newBuilderForType( 15687 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15688 Builder builder = new Builder(parent); 15689 return builder; 15690 } 15691 /** 15692 * Protobuf type {@code hadoop.hdfs.PrepareRecoveryResponseProto} 15693 */ 15694 public static final class Builder extends 15695 com.google.protobuf.GeneratedMessage.Builder<Builder> 15696 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProtoOrBuilder { 15697 public static final com.google.protobuf.Descriptors.Descriptor 15698 getDescriptor() { 15699 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor; 15700 } 15701 15702 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15703 internalGetFieldAccessorTable() { 15704 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable 15705 .ensureFieldAccessorsInitialized( 15706 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class); 15707 } 15708 15709 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.newBuilder() 15710 private Builder() { 15711 maybeForceBuilderInitialization(); 15712 } 15713 15714 private Builder( 15715 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15716 super(parent); 15717 maybeForceBuilderInitialization(); 15718 } 15719 private void maybeForceBuilderInitialization() { 15720 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 15721 getSegmentStateFieldBuilder(); 15722 } 15723 } 15724 private static Builder create() { 15725 return new Builder(); 15726 } 15727 15728 public Builder clear() { 15729 super.clear(); 15730 if (segmentStateBuilder_ == null) { 15731 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 15732 } else { 15733 segmentStateBuilder_.clear(); 15734 } 15735 bitField0_ = (bitField0_ & ~0x00000001); 15736 acceptedInEpoch_ = 0L; 15737 bitField0_ = (bitField0_ & ~0x00000002); 15738 lastWriterEpoch_ = 0L; 15739 bitField0_ = (bitField0_ & ~0x00000004); 15740 lastCommittedTxId_ = 0L; 15741 bitField0_ = (bitField0_ & ~0x00000008); 15742 return this; 15743 } 15744 15745 public Builder clone() { 15746 return create().mergeFrom(buildPartial()); 15747 } 15748 15749 public com.google.protobuf.Descriptors.Descriptor 15750 getDescriptorForType() { 15751 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor; 15752 } 15753 15754 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto getDefaultInstanceForType() { 15755 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(); 15756 } 15757 15758 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto build() { 15759 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = buildPartial(); 15760 if (!result.isInitialized()) { 15761 throw newUninitializedMessageException(result); 15762 } 15763 return result; 15764 } 15765 15766 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto buildPartial() { 15767 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto(this); 15768 int from_bitField0_ = bitField0_; 15769 int to_bitField0_ = 0; 15770 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 15771 to_bitField0_ |= 0x00000001; 15772 } 15773 if (segmentStateBuilder_ == null) { 15774 result.segmentState_ = segmentState_; 15775 } else { 15776 result.segmentState_ = segmentStateBuilder_.build(); 15777 } 15778 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 15779 to_bitField0_ |= 0x00000002; 15780 } 15781 result.acceptedInEpoch_ = acceptedInEpoch_; 15782 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 15783 to_bitField0_ |= 0x00000004; 15784 } 15785 result.lastWriterEpoch_ = lastWriterEpoch_; 15786 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 15787 to_bitField0_ |= 0x00000008; 15788 } 15789 result.lastCommittedTxId_ = lastCommittedTxId_; 15790 result.bitField0_ = to_bitField0_; 15791 onBuilt(); 15792 return result; 15793 } 15794 15795 public Builder mergeFrom(com.google.protobuf.Message other) { 15796 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) { 15797 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)other); 15798 } else { 15799 super.mergeFrom(other); 15800 return this; 15801 } 15802 } 15803 15804 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other) { 15805 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()) return this; 15806 if (other.hasSegmentState()) { 15807 mergeSegmentState(other.getSegmentState()); 15808 } 15809 if (other.hasAcceptedInEpoch()) { 15810 setAcceptedInEpoch(other.getAcceptedInEpoch()); 15811 } 15812 if (other.hasLastWriterEpoch()) { 15813 setLastWriterEpoch(other.getLastWriterEpoch()); 15814 } 15815 if (other.hasLastCommittedTxId()) { 15816 setLastCommittedTxId(other.getLastCommittedTxId()); 15817 } 15818 this.mergeUnknownFields(other.getUnknownFields()); 15819 return this; 15820 } 15821 15822 public final boolean isInitialized() { 15823 if (!hasLastWriterEpoch()) { 15824 15825 return false; 15826 } 15827 if (hasSegmentState()) { 15828 if (!getSegmentState().isInitialized()) { 15829 15830 return false; 15831 } 15832 } 15833 return true; 15834 } 15835 15836 public Builder mergeFrom( 15837 com.google.protobuf.CodedInputStream input, 15838 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15839 throws java.io.IOException { 15840 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parsedMessage = null; 15841 try { 15842 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 15843 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15844 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) e.getUnfinishedMessage(); 15845 throw e; 15846 } finally { 15847 if (parsedMessage != null) { 15848 mergeFrom(parsedMessage); 15849 } 15850 } 15851 return this; 15852 } 15853 private int bitField0_; 15854 15855 // optional .hadoop.hdfs.SegmentStateProto segmentState = 1; 15856 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 15857 private com.google.protobuf.SingleFieldBuilder< 15858 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_; 15859 /** 15860 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15861 */ 15862 public boolean hasSegmentState() { 15863 return ((bitField0_ & 0x00000001) == 0x00000001); 15864 } 15865 /** 15866 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15867 */ 15868 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() { 15869 if (segmentStateBuilder_ == null) { 15870 return segmentState_; 15871 } else { 15872 return segmentStateBuilder_.getMessage(); 15873 } 15874 } 15875 /** 15876 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15877 */ 15878 public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 15879 if (segmentStateBuilder_ == null) { 15880 if (value == null) { 15881 throw new NullPointerException(); 15882 } 15883 segmentState_ = value; 15884 onChanged(); 15885 } else { 15886 segmentStateBuilder_.setMessage(value); 15887 } 15888 bitField0_ |= 0x00000001; 15889 return this; 15890 } 15891 /** 15892 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15893 */ 15894 public Builder setSegmentState( 15895 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) { 15896 if (segmentStateBuilder_ == null) { 15897 segmentState_ = builderForValue.build(); 15898 onChanged(); 15899 } else { 15900 segmentStateBuilder_.setMessage(builderForValue.build()); 15901 } 15902 bitField0_ |= 0x00000001; 15903 return this; 15904 } 15905 /** 15906 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15907 */ 15908 public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 15909 if (segmentStateBuilder_ == null) { 15910 if (((bitField0_ & 0x00000001) == 0x00000001) && 15911 segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) { 15912 segmentState_ = 15913 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial(); 15914 } else { 15915 segmentState_ = value; 15916 } 15917 onChanged(); 15918 } else { 15919 segmentStateBuilder_.mergeFrom(value); 15920 } 15921 bitField0_ |= 0x00000001; 15922 return this; 15923 } 15924 /** 15925 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15926 */ 15927 public Builder clearSegmentState() { 15928 if (segmentStateBuilder_ == null) { 15929 segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 15930 onChanged(); 15931 } else { 15932 segmentStateBuilder_.clear(); 15933 } 15934 bitField0_ = (bitField0_ & ~0x00000001); 15935 return this; 15936 } 15937 /** 15938 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15939 */ 15940 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() { 15941 bitField0_ |= 0x00000001; 15942 onChanged(); 15943 return getSegmentStateFieldBuilder().getBuilder(); 15944 } 15945 /** 15946 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15947 */ 15948 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() { 15949 if (segmentStateBuilder_ != null) { 15950 return segmentStateBuilder_.getMessageOrBuilder(); 15951 } else { 15952 return segmentState_; 15953 } 15954 } 15955 /** 15956 * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code> 15957 */ 15958 private com.google.protobuf.SingleFieldBuilder< 15959 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 15960 getSegmentStateFieldBuilder() { 15961 if (segmentStateBuilder_ == null) { 15962 segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder< 15963 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>( 15964 segmentState_, 15965 getParentForChildren(), 15966 isClean()); 15967 segmentState_ = null; 15968 } 15969 return segmentStateBuilder_; 15970 } 15971 15972 // optional uint64 acceptedInEpoch = 2; 15973 private long acceptedInEpoch_ ; 15974 /** 15975 * <code>optional uint64 acceptedInEpoch = 2;</code> 15976 */ 15977 public boolean hasAcceptedInEpoch() { 15978 return ((bitField0_ & 0x00000002) == 0x00000002); 15979 } 15980 /** 15981 * <code>optional uint64 acceptedInEpoch = 2;</code> 15982 */ 15983 public long getAcceptedInEpoch() { 15984 return acceptedInEpoch_; 15985 } 15986 /** 15987 * <code>optional uint64 acceptedInEpoch = 2;</code> 15988 */ 15989 public Builder setAcceptedInEpoch(long value) { 15990 bitField0_ |= 0x00000002; 15991 acceptedInEpoch_ = value; 15992 onChanged(); 15993 return this; 15994 } 15995 /** 15996 * <code>optional uint64 acceptedInEpoch = 2;</code> 15997 */ 15998 public Builder clearAcceptedInEpoch() { 15999 bitField0_ = (bitField0_ & ~0x00000002); 16000 acceptedInEpoch_ = 0L; 16001 onChanged(); 16002 return this; 16003 } 16004 16005 // required uint64 lastWriterEpoch = 3; 16006 private long lastWriterEpoch_ ; 16007 /** 16008 * <code>required uint64 lastWriterEpoch = 3;</code> 16009 */ 16010 public boolean hasLastWriterEpoch() { 16011 return ((bitField0_ & 0x00000004) == 0x00000004); 16012 } 16013 /** 16014 * <code>required uint64 lastWriterEpoch = 3;</code> 16015 */ 16016 public long getLastWriterEpoch() { 16017 return lastWriterEpoch_; 16018 } 16019 /** 16020 * <code>required uint64 lastWriterEpoch = 3;</code> 16021 */ 16022 public Builder setLastWriterEpoch(long value) { 16023 bitField0_ |= 0x00000004; 16024 lastWriterEpoch_ = value; 16025 onChanged(); 16026 return this; 16027 } 16028 /** 16029 * <code>required uint64 lastWriterEpoch = 3;</code> 16030 */ 16031 public Builder clearLastWriterEpoch() { 16032 bitField0_ = (bitField0_ & ~0x00000004); 16033 lastWriterEpoch_ = 0L; 16034 onChanged(); 16035 return this; 16036 } 16037 16038 // optional uint64 lastCommittedTxId = 4; 16039 private long lastCommittedTxId_ ; 16040 /** 16041 * <code>optional uint64 lastCommittedTxId = 4;</code> 16042 * 16043 * <pre> 16044 * The highest committed txid that this logger has ever seen. 16045 * This may be higher than the data it actually has, in the case 16046 * that it was lagging before the old writer crashed. 16047 * </pre> 16048 */ 16049 public boolean hasLastCommittedTxId() { 16050 return ((bitField0_ & 0x00000008) == 0x00000008); 16051 } 16052 /** 16053 * <code>optional uint64 lastCommittedTxId = 4;</code> 16054 * 16055 * <pre> 16056 * The highest committed txid that this logger has ever seen. 16057 * This may be higher than the data it actually has, in the case 16058 * that it was lagging before the old writer crashed. 16059 * </pre> 16060 */ 16061 public long getLastCommittedTxId() { 16062 return lastCommittedTxId_; 16063 } 16064 /** 16065 * <code>optional uint64 lastCommittedTxId = 4;</code> 16066 * 16067 * <pre> 16068 * The highest committed txid that this logger has ever seen. 16069 * This may be higher than the data it actually has, in the case 16070 * that it was lagging before the old writer crashed. 16071 * </pre> 16072 */ 16073 public Builder setLastCommittedTxId(long value) { 16074 bitField0_ |= 0x00000008; 16075 lastCommittedTxId_ = value; 16076 onChanged(); 16077 return this; 16078 } 16079 /** 16080 * <code>optional uint64 lastCommittedTxId = 4;</code> 16081 * 16082 * <pre> 16083 * The highest committed txid that this logger has ever seen. 16084 * This may be higher than the data it actually has, in the case 16085 * that it was lagging before the old writer crashed. 16086 * </pre> 16087 */ 16088 public Builder clearLastCommittedTxId() { 16089 bitField0_ = (bitField0_ & ~0x00000008); 16090 lastCommittedTxId_ = 0L; 16091 onChanged(); 16092 return this; 16093 } 16094 16095 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryResponseProto) 16096 } 16097 16098 static { 16099 defaultInstance = new PrepareRecoveryResponseProto(true); 16100 defaultInstance.initFields(); 16101 } 16102 16103 // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryResponseProto) 16104 } 16105 16106 public interface AcceptRecoveryRequestProtoOrBuilder 16107 extends com.google.protobuf.MessageOrBuilder { 16108 16109 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 16110 /** 16111 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16112 */ 16113 boolean hasReqInfo(); 16114 /** 16115 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16116 */ 16117 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo(); 16118 /** 16119 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16120 */ 16121 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder(); 16122 16123 // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2; 16124 /** 16125 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16126 * 16127 * <pre> 16128 ** Details on the segment to recover 16129 * </pre> 16130 */ 16131 boolean hasStateToAccept(); 16132 /** 16133 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16134 * 16135 * <pre> 16136 ** Details on the segment to recover 16137 * </pre> 16138 */ 16139 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept(); 16140 /** 16141 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16142 * 16143 * <pre> 16144 ** Details on the segment to recover 16145 * </pre> 16146 */ 16147 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder(); 16148 16149 // required string fromURL = 3; 16150 /** 16151 * <code>required string fromURL = 3;</code> 16152 * 16153 * <pre> 16154 ** The URL from which the log may be copied 16155 * </pre> 16156 */ 16157 boolean hasFromURL(); 16158 /** 16159 * <code>required string fromURL = 3;</code> 16160 * 16161 * <pre> 16162 ** The URL from which the log may be copied 16163 * </pre> 16164 */ 16165 java.lang.String getFromURL(); 16166 /** 16167 * <code>required string fromURL = 3;</code> 16168 * 16169 * <pre> 16170 ** The URL from which the log may be copied 16171 * </pre> 16172 */ 16173 com.google.protobuf.ByteString 16174 getFromURLBytes(); 16175 } 16176 /** 16177 * Protobuf type {@code hadoop.hdfs.AcceptRecoveryRequestProto} 16178 * 16179 * <pre> 16180 ** 16181 * acceptRecovery() 16182 * </pre> 16183 */ 16184 public static final class AcceptRecoveryRequestProto extends 16185 com.google.protobuf.GeneratedMessage 16186 implements AcceptRecoveryRequestProtoOrBuilder { 16187 // Use AcceptRecoveryRequestProto.newBuilder() to construct. 16188 private AcceptRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 16189 super(builder); 16190 this.unknownFields = builder.getUnknownFields(); 16191 } 16192 private AcceptRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 16193 16194 private static final AcceptRecoveryRequestProto defaultInstance; 16195 public static AcceptRecoveryRequestProto getDefaultInstance() { 16196 return defaultInstance; 16197 } 16198 16199 public AcceptRecoveryRequestProto getDefaultInstanceForType() { 16200 return defaultInstance; 16201 } 16202 16203 private final com.google.protobuf.UnknownFieldSet unknownFields; 16204 @java.lang.Override 16205 public final com.google.protobuf.UnknownFieldSet 16206 getUnknownFields() { 16207 return this.unknownFields; 16208 } 16209 private AcceptRecoveryRequestProto( 16210 com.google.protobuf.CodedInputStream input, 16211 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16212 throws com.google.protobuf.InvalidProtocolBufferException { 16213 initFields(); 16214 int mutable_bitField0_ = 0; 16215 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 16216 com.google.protobuf.UnknownFieldSet.newBuilder(); 16217 try { 16218 boolean done = false; 16219 while (!done) { 16220 int tag = input.readTag(); 16221 switch (tag) { 16222 case 0: 16223 done = true; 16224 break; 16225 default: { 16226 if (!parseUnknownField(input, unknownFields, 16227 extensionRegistry, tag)) { 16228 done = true; 16229 } 16230 break; 16231 } 16232 case 10: { 16233 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null; 16234 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16235 subBuilder = reqInfo_.toBuilder(); 16236 } 16237 reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry); 16238 if (subBuilder != null) { 16239 subBuilder.mergeFrom(reqInfo_); 16240 reqInfo_ = subBuilder.buildPartial(); 16241 } 16242 bitField0_ |= 0x00000001; 16243 break; 16244 } 16245 case 18: { 16246 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null; 16247 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16248 subBuilder = stateToAccept_.toBuilder(); 16249 } 16250 stateToAccept_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry); 16251 if (subBuilder != null) { 16252 subBuilder.mergeFrom(stateToAccept_); 16253 stateToAccept_ = subBuilder.buildPartial(); 16254 } 16255 bitField0_ |= 0x00000002; 16256 break; 16257 } 16258 case 26: { 16259 bitField0_ |= 0x00000004; 16260 fromURL_ = input.readBytes(); 16261 break; 16262 } 16263 } 16264 } 16265 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16266 throw e.setUnfinishedMessage(this); 16267 } catch (java.io.IOException e) { 16268 throw new com.google.protobuf.InvalidProtocolBufferException( 16269 e.getMessage()).setUnfinishedMessage(this); 16270 } finally { 16271 this.unknownFields = unknownFields.build(); 16272 makeExtensionsImmutable(); 16273 } 16274 } 16275 public static final com.google.protobuf.Descriptors.Descriptor 16276 getDescriptor() { 16277 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor; 16278 } 16279 16280 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16281 internalGetFieldAccessorTable() { 16282 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable 16283 .ensureFieldAccessorsInitialized( 16284 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class); 16285 } 16286 16287 public static com.google.protobuf.Parser<AcceptRecoveryRequestProto> PARSER = 16288 new com.google.protobuf.AbstractParser<AcceptRecoveryRequestProto>() { 16289 public AcceptRecoveryRequestProto parsePartialFrom( 16290 com.google.protobuf.CodedInputStream input, 16291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16292 throws com.google.protobuf.InvalidProtocolBufferException { 16293 return new AcceptRecoveryRequestProto(input, extensionRegistry); 16294 } 16295 }; 16296 16297 @java.lang.Override 16298 public com.google.protobuf.Parser<AcceptRecoveryRequestProto> getParserForType() { 16299 return PARSER; 16300 } 16301 16302 private int bitField0_; 16303 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 16304 public static final int REQINFO_FIELD_NUMBER = 1; 16305 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_; 16306 /** 16307 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16308 */ 16309 public boolean hasReqInfo() { 16310 return ((bitField0_ & 0x00000001) == 0x00000001); 16311 } 16312 /** 16313 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16314 */ 16315 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 16316 return reqInfo_; 16317 } 16318 /** 16319 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16320 */ 16321 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 16322 return reqInfo_; 16323 } 16324 16325 // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2; 16326 public static final int STATETOACCEPT_FIELD_NUMBER = 2; 16327 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_; 16328 /** 16329 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16330 * 16331 * <pre> 16332 ** Details on the segment to recover 16333 * </pre> 16334 */ 16335 public boolean hasStateToAccept() { 16336 return ((bitField0_ & 0x00000002) == 0x00000002); 16337 } 16338 /** 16339 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16340 * 16341 * <pre> 16342 ** Details on the segment to recover 16343 * </pre> 16344 */ 16345 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() { 16346 return stateToAccept_; 16347 } 16348 /** 16349 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16350 * 16351 * <pre> 16352 ** Details on the segment to recover 16353 * </pre> 16354 */ 16355 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() { 16356 return stateToAccept_; 16357 } 16358 16359 // required string fromURL = 3; 16360 public static final int FROMURL_FIELD_NUMBER = 3; 16361 private java.lang.Object fromURL_; 16362 /** 16363 * <code>required string fromURL = 3;</code> 16364 * 16365 * <pre> 16366 ** The URL from which the log may be copied 16367 * </pre> 16368 */ 16369 public boolean hasFromURL() { 16370 return ((bitField0_ & 0x00000004) == 0x00000004); 16371 } 16372 /** 16373 * <code>required string fromURL = 3;</code> 16374 * 16375 * <pre> 16376 ** The URL from which the log may be copied 16377 * </pre> 16378 */ 16379 public java.lang.String getFromURL() { 16380 java.lang.Object ref = fromURL_; 16381 if (ref instanceof java.lang.String) { 16382 return (java.lang.String) ref; 16383 } else { 16384 com.google.protobuf.ByteString bs = 16385 (com.google.protobuf.ByteString) ref; 16386 java.lang.String s = bs.toStringUtf8(); 16387 if (bs.isValidUtf8()) { 16388 fromURL_ = s; 16389 } 16390 return s; 16391 } 16392 } 16393 /** 16394 * <code>required string fromURL = 3;</code> 16395 * 16396 * <pre> 16397 ** The URL from which the log may be copied 16398 * </pre> 16399 */ 16400 public com.google.protobuf.ByteString 16401 getFromURLBytes() { 16402 java.lang.Object ref = fromURL_; 16403 if (ref instanceof java.lang.String) { 16404 com.google.protobuf.ByteString b = 16405 com.google.protobuf.ByteString.copyFromUtf8( 16406 (java.lang.String) ref); 16407 fromURL_ = b; 16408 return b; 16409 } else { 16410 return (com.google.protobuf.ByteString) ref; 16411 } 16412 } 16413 16414 private void initFields() { 16415 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 16416 stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 16417 fromURL_ = ""; 16418 } 16419 private byte memoizedIsInitialized = -1; 16420 public final boolean isInitialized() { 16421 byte isInitialized = memoizedIsInitialized; 16422 if (isInitialized != -1) return isInitialized == 1; 16423 16424 if (!hasReqInfo()) { 16425 memoizedIsInitialized = 0; 16426 return false; 16427 } 16428 if (!hasStateToAccept()) { 16429 memoizedIsInitialized = 0; 16430 return false; 16431 } 16432 if (!hasFromURL()) { 16433 memoizedIsInitialized = 0; 16434 return false; 16435 } 16436 if (!getReqInfo().isInitialized()) { 16437 memoizedIsInitialized = 0; 16438 return false; 16439 } 16440 if (!getStateToAccept().isInitialized()) { 16441 memoizedIsInitialized = 0; 16442 return false; 16443 } 16444 memoizedIsInitialized = 1; 16445 return true; 16446 } 16447 16448 public void writeTo(com.google.protobuf.CodedOutputStream output) 16449 throws java.io.IOException { 16450 getSerializedSize(); 16451 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16452 output.writeMessage(1, reqInfo_); 16453 } 16454 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16455 output.writeMessage(2, stateToAccept_); 16456 } 16457 if (((bitField0_ & 0x00000004) == 0x00000004)) { 16458 output.writeBytes(3, getFromURLBytes()); 16459 } 16460 getUnknownFields().writeTo(output); 16461 } 16462 16463 private int memoizedSerializedSize = -1; 16464 public int getSerializedSize() { 16465 int size = memoizedSerializedSize; 16466 if (size != -1) return size; 16467 16468 size = 0; 16469 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16470 size += com.google.protobuf.CodedOutputStream 16471 .computeMessageSize(1, reqInfo_); 16472 } 16473 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16474 size += com.google.protobuf.CodedOutputStream 16475 .computeMessageSize(2, stateToAccept_); 16476 } 16477 if (((bitField0_ & 0x00000004) == 0x00000004)) { 16478 size += com.google.protobuf.CodedOutputStream 16479 .computeBytesSize(3, getFromURLBytes()); 16480 } 16481 size += getUnknownFields().getSerializedSize(); 16482 memoizedSerializedSize = size; 16483 return size; 16484 } 16485 16486 private static final long serialVersionUID = 0L; 16487 @java.lang.Override 16488 protected java.lang.Object writeReplace() 16489 throws java.io.ObjectStreamException { 16490 return super.writeReplace(); 16491 } 16492 16493 @java.lang.Override 16494 public boolean equals(final java.lang.Object obj) { 16495 if (obj == this) { 16496 return true; 16497 } 16498 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)) { 16499 return super.equals(obj); 16500 } 16501 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) obj; 16502 16503 boolean result = true; 16504 result = result && (hasReqInfo() == other.hasReqInfo()); 16505 if (hasReqInfo()) { 16506 result = result && getReqInfo() 16507 .equals(other.getReqInfo()); 16508 } 16509 result = result && (hasStateToAccept() == other.hasStateToAccept()); 16510 if (hasStateToAccept()) { 16511 result = result && getStateToAccept() 16512 .equals(other.getStateToAccept()); 16513 } 16514 result = result && (hasFromURL() == other.hasFromURL()); 16515 if (hasFromURL()) { 16516 result = result && getFromURL() 16517 .equals(other.getFromURL()); 16518 } 16519 result = result && 16520 getUnknownFields().equals(other.getUnknownFields()); 16521 return result; 16522 } 16523 16524 private int memoizedHashCode = 0; 16525 @java.lang.Override 16526 public int hashCode() { 16527 if (memoizedHashCode != 0) { 16528 return memoizedHashCode; 16529 } 16530 int hash = 41; 16531 hash = (19 * hash) + getDescriptorForType().hashCode(); 16532 if (hasReqInfo()) { 16533 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 16534 hash = (53 * hash) + getReqInfo().hashCode(); 16535 } 16536 if (hasStateToAccept()) { 16537 hash = (37 * hash) + STATETOACCEPT_FIELD_NUMBER; 16538 hash = (53 * hash) + getStateToAccept().hashCode(); 16539 } 16540 if (hasFromURL()) { 16541 hash = (37 * hash) + FROMURL_FIELD_NUMBER; 16542 hash = (53 * hash) + getFromURL().hashCode(); 16543 } 16544 hash = (29 * hash) + getUnknownFields().hashCode(); 16545 memoizedHashCode = hash; 16546 return hash; 16547 } 16548 16549 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16550 com.google.protobuf.ByteString data) 16551 throws com.google.protobuf.InvalidProtocolBufferException { 16552 return PARSER.parseFrom(data); 16553 } 16554 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16555 com.google.protobuf.ByteString data, 16556 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16557 throws com.google.protobuf.InvalidProtocolBufferException { 16558 return PARSER.parseFrom(data, extensionRegistry); 16559 } 16560 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(byte[] data) 16561 throws com.google.protobuf.InvalidProtocolBufferException { 16562 return PARSER.parseFrom(data); 16563 } 16564 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16565 byte[] data, 16566 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16567 throws com.google.protobuf.InvalidProtocolBufferException { 16568 return PARSER.parseFrom(data, extensionRegistry); 16569 } 16570 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(java.io.InputStream input) 16571 throws java.io.IOException { 16572 return PARSER.parseFrom(input); 16573 } 16574 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16575 java.io.InputStream input, 16576 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16577 throws java.io.IOException { 16578 return PARSER.parseFrom(input, extensionRegistry); 16579 } 16580 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input) 16581 throws java.io.IOException { 16582 return PARSER.parseDelimitedFrom(input); 16583 } 16584 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom( 16585 java.io.InputStream input, 16586 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16587 throws java.io.IOException { 16588 return PARSER.parseDelimitedFrom(input, extensionRegistry); 16589 } 16590 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16591 com.google.protobuf.CodedInputStream input) 16592 throws java.io.IOException { 16593 return PARSER.parseFrom(input); 16594 } 16595 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom( 16596 com.google.protobuf.CodedInputStream input, 16597 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16598 throws java.io.IOException { 16599 return PARSER.parseFrom(input, extensionRegistry); 16600 } 16601 16602 public static Builder newBuilder() { return Builder.create(); } 16603 public Builder newBuilderForType() { return newBuilder(); } 16604 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto prototype) { 16605 return newBuilder().mergeFrom(prototype); 16606 } 16607 public Builder toBuilder() { return newBuilder(this); } 16608 16609 @java.lang.Override 16610 protected Builder newBuilderForType( 16611 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16612 Builder builder = new Builder(parent); 16613 return builder; 16614 } 16615 /** 16616 * Protobuf type {@code hadoop.hdfs.AcceptRecoveryRequestProto} 16617 * 16618 * <pre> 16619 ** 16620 * acceptRecovery() 16621 * </pre> 16622 */ 16623 public static final class Builder extends 16624 com.google.protobuf.GeneratedMessage.Builder<Builder> 16625 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProtoOrBuilder { 16626 public static final com.google.protobuf.Descriptors.Descriptor 16627 getDescriptor() { 16628 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor; 16629 } 16630 16631 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16632 internalGetFieldAccessorTable() { 16633 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable 16634 .ensureFieldAccessorsInitialized( 16635 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class); 16636 } 16637 16638 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.newBuilder() 16639 private Builder() { 16640 maybeForceBuilderInitialization(); 16641 } 16642 16643 private Builder( 16644 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16645 super(parent); 16646 maybeForceBuilderInitialization(); 16647 } 16648 private void maybeForceBuilderInitialization() { 16649 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 16650 getReqInfoFieldBuilder(); 16651 getStateToAcceptFieldBuilder(); 16652 } 16653 } 16654 private static Builder create() { 16655 return new Builder(); 16656 } 16657 16658 public Builder clear() { 16659 super.clear(); 16660 if (reqInfoBuilder_ == null) { 16661 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 16662 } else { 16663 reqInfoBuilder_.clear(); 16664 } 16665 bitField0_ = (bitField0_ & ~0x00000001); 16666 if (stateToAcceptBuilder_ == null) { 16667 stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 16668 } else { 16669 stateToAcceptBuilder_.clear(); 16670 } 16671 bitField0_ = (bitField0_ & ~0x00000002); 16672 fromURL_ = ""; 16673 bitField0_ = (bitField0_ & ~0x00000004); 16674 return this; 16675 } 16676 16677 public Builder clone() { 16678 return create().mergeFrom(buildPartial()); 16679 } 16680 16681 public com.google.protobuf.Descriptors.Descriptor 16682 getDescriptorForType() { 16683 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor; 16684 } 16685 16686 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto getDefaultInstanceForType() { 16687 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance(); 16688 } 16689 16690 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto build() { 16691 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = buildPartial(); 16692 if (!result.isInitialized()) { 16693 throw newUninitializedMessageException(result); 16694 } 16695 return result; 16696 } 16697 16698 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto buildPartial() { 16699 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto(this); 16700 int from_bitField0_ = bitField0_; 16701 int to_bitField0_ = 0; 16702 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 16703 to_bitField0_ |= 0x00000001; 16704 } 16705 if (reqInfoBuilder_ == null) { 16706 result.reqInfo_ = reqInfo_; 16707 } else { 16708 result.reqInfo_ = reqInfoBuilder_.build(); 16709 } 16710 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 16711 to_bitField0_ |= 0x00000002; 16712 } 16713 if (stateToAcceptBuilder_ == null) { 16714 result.stateToAccept_ = stateToAccept_; 16715 } else { 16716 result.stateToAccept_ = stateToAcceptBuilder_.build(); 16717 } 16718 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 16719 to_bitField0_ |= 0x00000004; 16720 } 16721 result.fromURL_ = fromURL_; 16722 result.bitField0_ = to_bitField0_; 16723 onBuilt(); 16724 return result; 16725 } 16726 16727 public Builder mergeFrom(com.google.protobuf.Message other) { 16728 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) { 16729 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)other); 16730 } else { 16731 super.mergeFrom(other); 16732 return this; 16733 } 16734 } 16735 16736 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other) { 16737 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance()) return this; 16738 if (other.hasReqInfo()) { 16739 mergeReqInfo(other.getReqInfo()); 16740 } 16741 if (other.hasStateToAccept()) { 16742 mergeStateToAccept(other.getStateToAccept()); 16743 } 16744 if (other.hasFromURL()) { 16745 bitField0_ |= 0x00000004; 16746 fromURL_ = other.fromURL_; 16747 onChanged(); 16748 } 16749 this.mergeUnknownFields(other.getUnknownFields()); 16750 return this; 16751 } 16752 16753 public final boolean isInitialized() { 16754 if (!hasReqInfo()) { 16755 16756 return false; 16757 } 16758 if (!hasStateToAccept()) { 16759 16760 return false; 16761 } 16762 if (!hasFromURL()) { 16763 16764 return false; 16765 } 16766 if (!getReqInfo().isInitialized()) { 16767 16768 return false; 16769 } 16770 if (!getStateToAccept().isInitialized()) { 16771 16772 return false; 16773 } 16774 return true; 16775 } 16776 16777 public Builder mergeFrom( 16778 com.google.protobuf.CodedInputStream input, 16779 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16780 throws java.io.IOException { 16781 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parsedMessage = null; 16782 try { 16783 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 16784 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16785 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) e.getUnfinishedMessage(); 16786 throw e; 16787 } finally { 16788 if (parsedMessage != null) { 16789 mergeFrom(parsedMessage); 16790 } 16791 } 16792 return this; 16793 } 16794 private int bitField0_; 16795 16796 // required .hadoop.hdfs.RequestInfoProto reqInfo = 1; 16797 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 16798 private com.google.protobuf.SingleFieldBuilder< 16799 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_; 16800 /** 16801 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16802 */ 16803 public boolean hasReqInfo() { 16804 return ((bitField0_ & 0x00000001) == 0x00000001); 16805 } 16806 /** 16807 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16808 */ 16809 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() { 16810 if (reqInfoBuilder_ == null) { 16811 return reqInfo_; 16812 } else { 16813 return reqInfoBuilder_.getMessage(); 16814 } 16815 } 16816 /** 16817 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16818 */ 16819 public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 16820 if (reqInfoBuilder_ == null) { 16821 if (value == null) { 16822 throw new NullPointerException(); 16823 } 16824 reqInfo_ = value; 16825 onChanged(); 16826 } else { 16827 reqInfoBuilder_.setMessage(value); 16828 } 16829 bitField0_ |= 0x00000001; 16830 return this; 16831 } 16832 /** 16833 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16834 */ 16835 public Builder setReqInfo( 16836 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) { 16837 if (reqInfoBuilder_ == null) { 16838 reqInfo_ = builderForValue.build(); 16839 onChanged(); 16840 } else { 16841 reqInfoBuilder_.setMessage(builderForValue.build()); 16842 } 16843 bitField0_ |= 0x00000001; 16844 return this; 16845 } 16846 /** 16847 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16848 */ 16849 public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) { 16850 if (reqInfoBuilder_ == null) { 16851 if (((bitField0_ & 0x00000001) == 0x00000001) && 16852 reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) { 16853 reqInfo_ = 16854 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 16855 } else { 16856 reqInfo_ = value; 16857 } 16858 onChanged(); 16859 } else { 16860 reqInfoBuilder_.mergeFrom(value); 16861 } 16862 bitField0_ |= 0x00000001; 16863 return this; 16864 } 16865 /** 16866 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16867 */ 16868 public Builder clearReqInfo() { 16869 if (reqInfoBuilder_ == null) { 16870 reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance(); 16871 onChanged(); 16872 } else { 16873 reqInfoBuilder_.clear(); 16874 } 16875 bitField0_ = (bitField0_ & ~0x00000001); 16876 return this; 16877 } 16878 /** 16879 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16880 */ 16881 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() { 16882 bitField0_ |= 0x00000001; 16883 onChanged(); 16884 return getReqInfoFieldBuilder().getBuilder(); 16885 } 16886 /** 16887 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16888 */ 16889 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() { 16890 if (reqInfoBuilder_ != null) { 16891 return reqInfoBuilder_.getMessageOrBuilder(); 16892 } else { 16893 return reqInfo_; 16894 } 16895 } 16896 /** 16897 * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code> 16898 */ 16899 private com.google.protobuf.SingleFieldBuilder< 16900 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 16901 getReqInfoFieldBuilder() { 16902 if (reqInfoBuilder_ == null) { 16903 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 16904 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>( 16905 reqInfo_, 16906 getParentForChildren(), 16907 isClean()); 16908 reqInfo_ = null; 16909 } 16910 return reqInfoBuilder_; 16911 } 16912 16913 // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2; 16914 private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 16915 private com.google.protobuf.SingleFieldBuilder< 16916 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> stateToAcceptBuilder_; 16917 /** 16918 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16919 * 16920 * <pre> 16921 ** Details on the segment to recover 16922 * </pre> 16923 */ 16924 public boolean hasStateToAccept() { 16925 return ((bitField0_ & 0x00000002) == 0x00000002); 16926 } 16927 /** 16928 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16929 * 16930 * <pre> 16931 ** Details on the segment to recover 16932 * </pre> 16933 */ 16934 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() { 16935 if (stateToAcceptBuilder_ == null) { 16936 return stateToAccept_; 16937 } else { 16938 return stateToAcceptBuilder_.getMessage(); 16939 } 16940 } 16941 /** 16942 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16943 * 16944 * <pre> 16945 ** Details on the segment to recover 16946 * </pre> 16947 */ 16948 public Builder setStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 16949 if (stateToAcceptBuilder_ == null) { 16950 if (value == null) { 16951 throw new NullPointerException(); 16952 } 16953 stateToAccept_ = value; 16954 onChanged(); 16955 } else { 16956 stateToAcceptBuilder_.setMessage(value); 16957 } 16958 bitField0_ |= 0x00000002; 16959 return this; 16960 } 16961 /** 16962 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16963 * 16964 * <pre> 16965 ** Details on the segment to recover 16966 * </pre> 16967 */ 16968 public Builder setStateToAccept( 16969 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) { 16970 if (stateToAcceptBuilder_ == null) { 16971 stateToAccept_ = builderForValue.build(); 16972 onChanged(); 16973 } else { 16974 stateToAcceptBuilder_.setMessage(builderForValue.build()); 16975 } 16976 bitField0_ |= 0x00000002; 16977 return this; 16978 } 16979 /** 16980 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 16981 * 16982 * <pre> 16983 ** Details on the segment to recover 16984 * </pre> 16985 */ 16986 public Builder mergeStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) { 16987 if (stateToAcceptBuilder_ == null) { 16988 if (((bitField0_ & 0x00000002) == 0x00000002) && 16989 stateToAccept_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) { 16990 stateToAccept_ = 16991 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(stateToAccept_).mergeFrom(value).buildPartial(); 16992 } else { 16993 stateToAccept_ = value; 16994 } 16995 onChanged(); 16996 } else { 16997 stateToAcceptBuilder_.mergeFrom(value); 16998 } 16999 bitField0_ |= 0x00000002; 17000 return this; 17001 } 17002 /** 17003 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 17004 * 17005 * <pre> 17006 ** Details on the segment to recover 17007 * </pre> 17008 */ 17009 public Builder clearStateToAccept() { 17010 if (stateToAcceptBuilder_ == null) { 17011 stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance(); 17012 onChanged(); 17013 } else { 17014 stateToAcceptBuilder_.clear(); 17015 } 17016 bitField0_ = (bitField0_ & ~0x00000002); 17017 return this; 17018 } 17019 /** 17020 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 17021 * 17022 * <pre> 17023 ** Details on the segment to recover 17024 * </pre> 17025 */ 17026 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getStateToAcceptBuilder() { 17027 bitField0_ |= 0x00000002; 17028 onChanged(); 17029 return getStateToAcceptFieldBuilder().getBuilder(); 17030 } 17031 /** 17032 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 17033 * 17034 * <pre> 17035 ** Details on the segment to recover 17036 * </pre> 17037 */ 17038 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() { 17039 if (stateToAcceptBuilder_ != null) { 17040 return stateToAcceptBuilder_.getMessageOrBuilder(); 17041 } else { 17042 return stateToAccept_; 17043 } 17044 } 17045 /** 17046 * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code> 17047 * 17048 * <pre> 17049 ** Details on the segment to recover 17050 * </pre> 17051 */ 17052 private com.google.protobuf.SingleFieldBuilder< 17053 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 17054 getStateToAcceptFieldBuilder() { 17055 if (stateToAcceptBuilder_ == null) { 17056 stateToAcceptBuilder_ = new com.google.protobuf.SingleFieldBuilder< 17057 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>( 17058 stateToAccept_, 17059 getParentForChildren(), 17060 isClean()); 17061 stateToAccept_ = null; 17062 } 17063 return stateToAcceptBuilder_; 17064 } 17065 17066 // required string fromURL = 3; 17067 private java.lang.Object fromURL_ = ""; 17068 /** 17069 * <code>required string fromURL = 3;</code> 17070 * 17071 * <pre> 17072 ** The URL from which the log may be copied 17073 * </pre> 17074 */ 17075 public boolean hasFromURL() { 17076 return ((bitField0_ & 0x00000004) == 0x00000004); 17077 } 17078 /** 17079 * <code>required string fromURL = 3;</code> 17080 * 17081 * <pre> 17082 ** The URL from which the log may be copied 17083 * </pre> 17084 */ 17085 public java.lang.String getFromURL() { 17086 java.lang.Object ref = fromURL_; 17087 if (!(ref instanceof java.lang.String)) { 17088 java.lang.String s = ((com.google.protobuf.ByteString) ref) 17089 .toStringUtf8(); 17090 fromURL_ = s; 17091 return s; 17092 } else { 17093 return (java.lang.String) ref; 17094 } 17095 } 17096 /** 17097 * <code>required string fromURL = 3;</code> 17098 * 17099 * <pre> 17100 ** The URL from which the log may be copied 17101 * </pre> 17102 */ 17103 public com.google.protobuf.ByteString 17104 getFromURLBytes() { 17105 java.lang.Object ref = fromURL_; 17106 if (ref instanceof String) { 17107 com.google.protobuf.ByteString b = 17108 com.google.protobuf.ByteString.copyFromUtf8( 17109 (java.lang.String) ref); 17110 fromURL_ = b; 17111 return b; 17112 } else { 17113 return (com.google.protobuf.ByteString) ref; 17114 } 17115 } 17116 /** 17117 * <code>required string fromURL = 3;</code> 17118 * 17119 * <pre> 17120 ** The URL from which the log may be copied 17121 * </pre> 17122 */ 17123 public Builder setFromURL( 17124 java.lang.String value) { 17125 if (value == null) { 17126 throw new NullPointerException(); 17127 } 17128 bitField0_ |= 0x00000004; 17129 fromURL_ = value; 17130 onChanged(); 17131 return this; 17132 } 17133 /** 17134 * <code>required string fromURL = 3;</code> 17135 * 17136 * <pre> 17137 ** The URL from which the log may be copied 17138 * </pre> 17139 */ 17140 public Builder clearFromURL() { 17141 bitField0_ = (bitField0_ & ~0x00000004); 17142 fromURL_ = getDefaultInstance().getFromURL(); 17143 onChanged(); 17144 return this; 17145 } 17146 /** 17147 * <code>required string fromURL = 3;</code> 17148 * 17149 * <pre> 17150 ** The URL from which the log may be copied 17151 * </pre> 17152 */ 17153 public Builder setFromURLBytes( 17154 com.google.protobuf.ByteString value) { 17155 if (value == null) { 17156 throw new NullPointerException(); 17157 } 17158 bitField0_ |= 0x00000004; 17159 fromURL_ = value; 17160 onChanged(); 17161 return this; 17162 } 17163 17164 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryRequestProto) 17165 } 17166 17167 static { 17168 defaultInstance = new AcceptRecoveryRequestProto(true); 17169 defaultInstance.initFields(); 17170 } 17171 17172 // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryRequestProto) 17173 } 17174 17175 public interface AcceptRecoveryResponseProtoOrBuilder 17176 extends com.google.protobuf.MessageOrBuilder { 17177 } 17178 /** 17179 * Protobuf type {@code hadoop.hdfs.AcceptRecoveryResponseProto} 17180 */ 17181 public static final class AcceptRecoveryResponseProto extends 17182 com.google.protobuf.GeneratedMessage 17183 implements AcceptRecoveryResponseProtoOrBuilder { 17184 // Use AcceptRecoveryResponseProto.newBuilder() to construct. 17185 private AcceptRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 17186 super(builder); 17187 this.unknownFields = builder.getUnknownFields(); 17188 } 17189 private AcceptRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 17190 17191 private static final AcceptRecoveryResponseProto defaultInstance; 17192 public static AcceptRecoveryResponseProto getDefaultInstance() { 17193 return defaultInstance; 17194 } 17195 17196 public AcceptRecoveryResponseProto getDefaultInstanceForType() { 17197 return defaultInstance; 17198 } 17199 17200 private final com.google.protobuf.UnknownFieldSet unknownFields; 17201 @java.lang.Override 17202 public final com.google.protobuf.UnknownFieldSet 17203 getUnknownFields() { 17204 return this.unknownFields; 17205 } 17206 private AcceptRecoveryResponseProto( 17207 com.google.protobuf.CodedInputStream input, 17208 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17209 throws com.google.protobuf.InvalidProtocolBufferException { 17210 initFields(); 17211 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 17212 com.google.protobuf.UnknownFieldSet.newBuilder(); 17213 try { 17214 boolean done = false; 17215 while (!done) { 17216 int tag = input.readTag(); 17217 switch (tag) { 17218 case 0: 17219 done = true; 17220 break; 17221 default: { 17222 if (!parseUnknownField(input, unknownFields, 17223 extensionRegistry, tag)) { 17224 done = true; 17225 } 17226 break; 17227 } 17228 } 17229 } 17230 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17231 throw e.setUnfinishedMessage(this); 17232 } catch (java.io.IOException e) { 17233 throw new com.google.protobuf.InvalidProtocolBufferException( 17234 e.getMessage()).setUnfinishedMessage(this); 17235 } finally { 17236 this.unknownFields = unknownFields.build(); 17237 makeExtensionsImmutable(); 17238 } 17239 } 17240 public static final com.google.protobuf.Descriptors.Descriptor 17241 getDescriptor() { 17242 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor; 17243 } 17244 17245 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17246 internalGetFieldAccessorTable() { 17247 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable 17248 .ensureFieldAccessorsInitialized( 17249 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class); 17250 } 17251 17252 public static com.google.protobuf.Parser<AcceptRecoveryResponseProto> PARSER = 17253 new com.google.protobuf.AbstractParser<AcceptRecoveryResponseProto>() { 17254 public AcceptRecoveryResponseProto parsePartialFrom( 17255 com.google.protobuf.CodedInputStream input, 17256 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17257 throws com.google.protobuf.InvalidProtocolBufferException { 17258 return new AcceptRecoveryResponseProto(input, extensionRegistry); 17259 } 17260 }; 17261 17262 @java.lang.Override 17263 public com.google.protobuf.Parser<AcceptRecoveryResponseProto> getParserForType() { 17264 return PARSER; 17265 } 17266 17267 private void initFields() { 17268 } 17269 private byte memoizedIsInitialized = -1; 17270 public final boolean isInitialized() { 17271 byte isInitialized = memoizedIsInitialized; 17272 if (isInitialized != -1) return isInitialized == 1; 17273 17274 memoizedIsInitialized = 1; 17275 return true; 17276 } 17277 17278 public void writeTo(com.google.protobuf.CodedOutputStream output) 17279 throws java.io.IOException { 17280 getSerializedSize(); 17281 getUnknownFields().writeTo(output); 17282 } 17283 17284 private int memoizedSerializedSize = -1; 17285 public int getSerializedSize() { 17286 int size = memoizedSerializedSize; 17287 if (size != -1) return size; 17288 17289 size = 0; 17290 size += getUnknownFields().getSerializedSize(); 17291 memoizedSerializedSize = size; 17292 return size; 17293 } 17294 17295 private static final long serialVersionUID = 0L; 17296 @java.lang.Override 17297 protected java.lang.Object writeReplace() 17298 throws java.io.ObjectStreamException { 17299 return super.writeReplace(); 17300 } 17301 17302 @java.lang.Override 17303 public boolean equals(final java.lang.Object obj) { 17304 if (obj == this) { 17305 return true; 17306 } 17307 if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)) { 17308 return super.equals(obj); 17309 } 17310 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) obj; 17311 17312 boolean result = true; 17313 result = result && 17314 getUnknownFields().equals(other.getUnknownFields()); 17315 return result; 17316 } 17317 17318 private int memoizedHashCode = 0; 17319 @java.lang.Override 17320 public int hashCode() { 17321 if (memoizedHashCode != 0) { 17322 return memoizedHashCode; 17323 } 17324 int hash = 41; 17325 hash = (19 * hash) + getDescriptorForType().hashCode(); 17326 hash = (29 * hash) + getUnknownFields().hashCode(); 17327 memoizedHashCode = hash; 17328 return hash; 17329 } 17330 17331 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17332 com.google.protobuf.ByteString data) 17333 throws com.google.protobuf.InvalidProtocolBufferException { 17334 return PARSER.parseFrom(data); 17335 } 17336 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17337 com.google.protobuf.ByteString data, 17338 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17339 throws com.google.protobuf.InvalidProtocolBufferException { 17340 return PARSER.parseFrom(data, extensionRegistry); 17341 } 17342 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(byte[] data) 17343 throws com.google.protobuf.InvalidProtocolBufferException { 17344 return PARSER.parseFrom(data); 17345 } 17346 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17347 byte[] data, 17348 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17349 throws com.google.protobuf.InvalidProtocolBufferException { 17350 return PARSER.parseFrom(data, extensionRegistry); 17351 } 17352 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(java.io.InputStream input) 17353 throws java.io.IOException { 17354 return PARSER.parseFrom(input); 17355 } 17356 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17357 java.io.InputStream input, 17358 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17359 throws java.io.IOException { 17360 return PARSER.parseFrom(input, extensionRegistry); 17361 } 17362 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input) 17363 throws java.io.IOException { 17364 return PARSER.parseDelimitedFrom(input); 17365 } 17366 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom( 17367 java.io.InputStream input, 17368 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17369 throws java.io.IOException { 17370 return PARSER.parseDelimitedFrom(input, extensionRegistry); 17371 } 17372 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17373 com.google.protobuf.CodedInputStream input) 17374 throws java.io.IOException { 17375 return PARSER.parseFrom(input); 17376 } 17377 public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom( 17378 com.google.protobuf.CodedInputStream input, 17379 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17380 throws java.io.IOException { 17381 return PARSER.parseFrom(input, extensionRegistry); 17382 } 17383 17384 public static Builder newBuilder() { return Builder.create(); } 17385 public Builder newBuilderForType() { return newBuilder(); } 17386 public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto prototype) { 17387 return newBuilder().mergeFrom(prototype); 17388 } 17389 public Builder toBuilder() { return newBuilder(this); } 17390 17391 @java.lang.Override 17392 protected Builder newBuilderForType( 17393 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17394 Builder builder = new Builder(parent); 17395 return builder; 17396 } 17397 /** 17398 * Protobuf type {@code hadoop.hdfs.AcceptRecoveryResponseProto} 17399 */ 17400 public static final class Builder extends 17401 com.google.protobuf.GeneratedMessage.Builder<Builder> 17402 implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProtoOrBuilder { 17403 public static final com.google.protobuf.Descriptors.Descriptor 17404 getDescriptor() { 17405 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor; 17406 } 17407 17408 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17409 internalGetFieldAccessorTable() { 17410 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable 17411 .ensureFieldAccessorsInitialized( 17412 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class); 17413 } 17414 17415 // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.newBuilder() 17416 private Builder() { 17417 maybeForceBuilderInitialization(); 17418 } 17419 17420 private Builder( 17421 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17422 super(parent); 17423 maybeForceBuilderInitialization(); 17424 } 17425 private void maybeForceBuilderInitialization() { 17426 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 17427 } 17428 } 17429 private static Builder create() { 17430 return new Builder(); 17431 } 17432 17433 public Builder clear() { 17434 super.clear(); 17435 return this; 17436 } 17437 17438 public Builder clone() { 17439 return create().mergeFrom(buildPartial()); 17440 } 17441 17442 public com.google.protobuf.Descriptors.Descriptor 17443 getDescriptorForType() { 17444 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor; 17445 } 17446 17447 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto getDefaultInstanceForType() { 17448 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(); 17449 } 17450 17451 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto build() { 17452 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = buildPartial(); 17453 if (!result.isInitialized()) { 17454 throw newUninitializedMessageException(result); 17455 } 17456 return result; 17457 } 17458 17459 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto buildPartial() { 17460 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto(this); 17461 onBuilt(); 17462 return result; 17463 } 17464 17465 public Builder mergeFrom(com.google.protobuf.Message other) { 17466 if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) { 17467 return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)other); 17468 } else { 17469 super.mergeFrom(other); 17470 return this; 17471 } 17472 } 17473 17474 public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other) { 17475 if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()) return this; 17476 this.mergeUnknownFields(other.getUnknownFields()); 17477 return this; 17478 } 17479 17480 public final boolean isInitialized() { 17481 return true; 17482 } 17483 17484 public Builder mergeFrom( 17485 com.google.protobuf.CodedInputStream input, 17486 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17487 throws java.io.IOException { 17488 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parsedMessage = null; 17489 try { 17490 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 17491 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17492 parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) e.getUnfinishedMessage(); 17493 throw e; 17494 } finally { 17495 if (parsedMessage != null) { 17496 mergeFrom(parsedMessage); 17497 } 17498 } 17499 return this; 17500 } 17501 17502 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryResponseProto) 17503 } 17504 17505 static { 17506 defaultInstance = new AcceptRecoveryResponseProto(true); 17507 defaultInstance.initFields(); 17508 } 17509 17510 // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryResponseProto) 17511 } 17512 17513 /** 17514 * Protobuf service {@code hadoop.hdfs.QJournalProtocolService} 17515 * 17516 * <pre> 17517 ** 17518 * Protocol used to journal edits to a JournalNode. 17519 * See the request and response for details of rpc call. 17520 * </pre> 17521 */ 17522 public static abstract class QJournalProtocolService 17523 implements com.google.protobuf.Service { 17524 protected QJournalProtocolService() {} 17525 17526 public interface Interface { 17527 /** 17528 * <code>rpc isFormatted(.hadoop.hdfs.IsFormattedRequestProto) returns (.hadoop.hdfs.IsFormattedResponseProto);</code> 17529 */ 17530 public abstract void isFormatted( 17531 com.google.protobuf.RpcController controller, 17532 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request, 17533 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done); 17534 17535 /** 17536 * <code>rpc getJournalState(.hadoop.hdfs.GetJournalStateRequestProto) returns (.hadoop.hdfs.GetJournalStateResponseProto);</code> 17537 */ 17538 public abstract void getJournalState( 17539 com.google.protobuf.RpcController controller, 17540 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request, 17541 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done); 17542 17543 /** 17544 * <code>rpc newEpoch(.hadoop.hdfs.NewEpochRequestProto) returns (.hadoop.hdfs.NewEpochResponseProto);</code> 17545 */ 17546 public abstract void newEpoch( 17547 com.google.protobuf.RpcController controller, 17548 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request, 17549 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done); 17550 17551 /** 17552 * <code>rpc format(.hadoop.hdfs.FormatRequestProto) returns (.hadoop.hdfs.FormatResponseProto);</code> 17553 */ 17554 public abstract void format( 17555 com.google.protobuf.RpcController controller, 17556 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request, 17557 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done); 17558 17559 /** 17560 * <code>rpc journal(.hadoop.hdfs.JournalRequestProto) returns (.hadoop.hdfs.JournalResponseProto);</code> 17561 */ 17562 public abstract void journal( 17563 com.google.protobuf.RpcController controller, 17564 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request, 17565 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done); 17566 17567 /** 17568 * <code>rpc heartbeat(.hadoop.hdfs.HeartbeatRequestProto) returns (.hadoop.hdfs.HeartbeatResponseProto);</code> 17569 */ 17570 public abstract void heartbeat( 17571 com.google.protobuf.RpcController controller, 17572 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request, 17573 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done); 17574 17575 /** 17576 * <code>rpc startLogSegment(.hadoop.hdfs.StartLogSegmentRequestProto) returns (.hadoop.hdfs.StartLogSegmentResponseProto);</code> 17577 */ 17578 public abstract void startLogSegment( 17579 com.google.protobuf.RpcController controller, 17580 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request, 17581 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done); 17582 17583 /** 17584 * <code>rpc finalizeLogSegment(.hadoop.hdfs.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.FinalizeLogSegmentResponseProto);</code> 17585 */ 17586 public abstract void finalizeLogSegment( 17587 com.google.protobuf.RpcController controller, 17588 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request, 17589 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done); 17590 17591 /** 17592 * <code>rpc purgeLogs(.hadoop.hdfs.PurgeLogsRequestProto) returns (.hadoop.hdfs.PurgeLogsResponseProto);</code> 17593 */ 17594 public abstract void purgeLogs( 17595 com.google.protobuf.RpcController controller, 17596 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request, 17597 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done); 17598 17599 /** 17600 * <code>rpc getEditLogManifest(.hadoop.hdfs.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.GetEditLogManifestResponseProto);</code> 17601 */ 17602 public abstract void getEditLogManifest( 17603 com.google.protobuf.RpcController controller, 17604 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request, 17605 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done); 17606 17607 /** 17608 * <code>rpc prepareRecovery(.hadoop.hdfs.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.PrepareRecoveryResponseProto);</code> 17609 */ 17610 public abstract void prepareRecovery( 17611 com.google.protobuf.RpcController controller, 17612 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request, 17613 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done); 17614 17615 /** 17616 * <code>rpc acceptRecovery(.hadoop.hdfs.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.AcceptRecoveryResponseProto);</code> 17617 */ 17618 public abstract void acceptRecovery( 17619 com.google.protobuf.RpcController controller, 17620 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request, 17621 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done); 17622 17623 } 17624 17625 public static com.google.protobuf.Service newReflectiveService( 17626 final Interface impl) { 17627 return new QJournalProtocolService() { 17628 @java.lang.Override 17629 public void isFormatted( 17630 com.google.protobuf.RpcController controller, 17631 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request, 17632 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) { 17633 impl.isFormatted(controller, request, done); 17634 } 17635 17636 @java.lang.Override 17637 public void getJournalState( 17638 com.google.protobuf.RpcController controller, 17639 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request, 17640 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) { 17641 impl.getJournalState(controller, request, done); 17642 } 17643 17644 @java.lang.Override 17645 public void newEpoch( 17646 com.google.protobuf.RpcController controller, 17647 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request, 17648 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) { 17649 impl.newEpoch(controller, request, done); 17650 } 17651 17652 @java.lang.Override 17653 public void format( 17654 com.google.protobuf.RpcController controller, 17655 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request, 17656 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) { 17657 impl.format(controller, request, done); 17658 } 17659 17660 @java.lang.Override 17661 public void journal( 17662 com.google.protobuf.RpcController controller, 17663 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request, 17664 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) { 17665 impl.journal(controller, request, done); 17666 } 17667 17668 @java.lang.Override 17669 public void heartbeat( 17670 com.google.protobuf.RpcController controller, 17671 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request, 17672 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) { 17673 impl.heartbeat(controller, request, done); 17674 } 17675 17676 @java.lang.Override 17677 public void startLogSegment( 17678 com.google.protobuf.RpcController controller, 17679 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request, 17680 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) { 17681 impl.startLogSegment(controller, request, done); 17682 } 17683 17684 @java.lang.Override 17685 public void finalizeLogSegment( 17686 com.google.protobuf.RpcController controller, 17687 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request, 17688 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) { 17689 impl.finalizeLogSegment(controller, request, done); 17690 } 17691 17692 @java.lang.Override 17693 public void purgeLogs( 17694 com.google.protobuf.RpcController controller, 17695 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request, 17696 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) { 17697 impl.purgeLogs(controller, request, done); 17698 } 17699 17700 @java.lang.Override 17701 public void getEditLogManifest( 17702 com.google.protobuf.RpcController controller, 17703 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request, 17704 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) { 17705 impl.getEditLogManifest(controller, request, done); 17706 } 17707 17708 @java.lang.Override 17709 public void prepareRecovery( 17710 com.google.protobuf.RpcController controller, 17711 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request, 17712 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) { 17713 impl.prepareRecovery(controller, request, done); 17714 } 17715 17716 @java.lang.Override 17717 public void acceptRecovery( 17718 com.google.protobuf.RpcController controller, 17719 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request, 17720 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) { 17721 impl.acceptRecovery(controller, request, done); 17722 } 17723 17724 }; 17725 } 17726 17727 public static com.google.protobuf.BlockingService 17728 newReflectiveBlockingService(final BlockingInterface impl) { 17729 return new com.google.protobuf.BlockingService() { 17730 public final com.google.protobuf.Descriptors.ServiceDescriptor 17731 getDescriptorForType() { 17732 return getDescriptor(); 17733 } 17734 17735 public final com.google.protobuf.Message callBlockingMethod( 17736 com.google.protobuf.Descriptors.MethodDescriptor method, 17737 com.google.protobuf.RpcController controller, 17738 com.google.protobuf.Message request) 17739 throws com.google.protobuf.ServiceException { 17740 if (method.getService() != getDescriptor()) { 17741 throw new java.lang.IllegalArgumentException( 17742 "Service.callBlockingMethod() given method descriptor for " + 17743 "wrong service type."); 17744 } 17745 switch(method.getIndex()) { 17746 case 0: 17747 return impl.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request); 17748 case 1: 17749 return impl.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request); 17750 case 2: 17751 return impl.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request); 17752 case 3: 17753 return impl.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request); 17754 case 4: 17755 return impl.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request); 17756 case 5: 17757 return impl.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request); 17758 case 6: 17759 return impl.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request); 17760 case 7: 17761 return impl.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request); 17762 case 8: 17763 return impl.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request); 17764 case 9: 17765 return impl.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request); 17766 case 10: 17767 return impl.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request); 17768 case 11: 17769 return impl.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request); 17770 default: 17771 throw new java.lang.AssertionError("Can't get here."); 17772 } 17773 } 17774 17775 public final com.google.protobuf.Message 17776 getRequestPrototype( 17777 com.google.protobuf.Descriptors.MethodDescriptor method) { 17778 if (method.getService() != getDescriptor()) { 17779 throw new java.lang.IllegalArgumentException( 17780 "Service.getRequestPrototype() given method " + 17781 "descriptor for wrong service type."); 17782 } 17783 switch(method.getIndex()) { 17784 case 0: 17785 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance(); 17786 case 1: 17787 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance(); 17788 case 2: 17789 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance(); 17790 case 3: 17791 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance(); 17792 case 4: 17793 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance(); 17794 case 5: 17795 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); 17796 case 6: 17797 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance(); 17798 case 7: 17799 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance(); 17800 case 8: 17801 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance(); 17802 case 9: 17803 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance(); 17804 case 10: 17805 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance(); 17806 case 11: 17807 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance(); 17808 default: 17809 throw new java.lang.AssertionError("Can't get here."); 17810 } 17811 } 17812 17813 public final com.google.protobuf.Message 17814 getResponsePrototype( 17815 com.google.protobuf.Descriptors.MethodDescriptor method) { 17816 if (method.getService() != getDescriptor()) { 17817 throw new java.lang.IllegalArgumentException( 17818 "Service.getResponsePrototype() given method " + 17819 "descriptor for wrong service type."); 17820 } 17821 switch(method.getIndex()) { 17822 case 0: 17823 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(); 17824 case 1: 17825 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(); 17826 case 2: 17827 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(); 17828 case 3: 17829 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(); 17830 case 4: 17831 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(); 17832 case 5: 17833 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); 17834 case 6: 17835 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(); 17836 case 7: 17837 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(); 17838 case 8: 17839 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(); 17840 case 9: 17841 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(); 17842 case 10: 17843 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(); 17844 case 11: 17845 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(); 17846 default: 17847 throw new java.lang.AssertionError("Can't get here."); 17848 } 17849 } 17850 17851 }; 17852 } 17853 17854 /** 17855 * <code>rpc isFormatted(.hadoop.hdfs.IsFormattedRequestProto) returns (.hadoop.hdfs.IsFormattedResponseProto);</code> 17856 */ 17857 public abstract void isFormatted( 17858 com.google.protobuf.RpcController controller, 17859 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request, 17860 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done); 17861 17862 /** 17863 * <code>rpc getJournalState(.hadoop.hdfs.GetJournalStateRequestProto) returns (.hadoop.hdfs.GetJournalStateResponseProto);</code> 17864 */ 17865 public abstract void getJournalState( 17866 com.google.protobuf.RpcController controller, 17867 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request, 17868 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done); 17869 17870 /** 17871 * <code>rpc newEpoch(.hadoop.hdfs.NewEpochRequestProto) returns (.hadoop.hdfs.NewEpochResponseProto);</code> 17872 */ 17873 public abstract void newEpoch( 17874 com.google.protobuf.RpcController controller, 17875 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request, 17876 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done); 17877 17878 /** 17879 * <code>rpc format(.hadoop.hdfs.FormatRequestProto) returns (.hadoop.hdfs.FormatResponseProto);</code> 17880 */ 17881 public abstract void format( 17882 com.google.protobuf.RpcController controller, 17883 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request, 17884 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done); 17885 17886 /** 17887 * <code>rpc journal(.hadoop.hdfs.JournalRequestProto) returns (.hadoop.hdfs.JournalResponseProto);</code> 17888 */ 17889 public abstract void journal( 17890 com.google.protobuf.RpcController controller, 17891 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request, 17892 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done); 17893 17894 /** 17895 * <code>rpc heartbeat(.hadoop.hdfs.HeartbeatRequestProto) returns (.hadoop.hdfs.HeartbeatResponseProto);</code> 17896 */ 17897 public abstract void heartbeat( 17898 com.google.protobuf.RpcController controller, 17899 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request, 17900 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done); 17901 17902 /** 17903 * <code>rpc startLogSegment(.hadoop.hdfs.StartLogSegmentRequestProto) returns (.hadoop.hdfs.StartLogSegmentResponseProto);</code> 17904 */ 17905 public abstract void startLogSegment( 17906 com.google.protobuf.RpcController controller, 17907 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request, 17908 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done); 17909 17910 /** 17911 * <code>rpc finalizeLogSegment(.hadoop.hdfs.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.FinalizeLogSegmentResponseProto);</code> 17912 */ 17913 public abstract void finalizeLogSegment( 17914 com.google.protobuf.RpcController controller, 17915 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request, 17916 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done); 17917 17918 /** 17919 * <code>rpc purgeLogs(.hadoop.hdfs.PurgeLogsRequestProto) returns (.hadoop.hdfs.PurgeLogsResponseProto);</code> 17920 */ 17921 public abstract void purgeLogs( 17922 com.google.protobuf.RpcController controller, 17923 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request, 17924 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done); 17925 17926 /** 17927 * <code>rpc getEditLogManifest(.hadoop.hdfs.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.GetEditLogManifestResponseProto);</code> 17928 */ 17929 public abstract void getEditLogManifest( 17930 com.google.protobuf.RpcController controller, 17931 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request, 17932 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done); 17933 17934 /** 17935 * <code>rpc prepareRecovery(.hadoop.hdfs.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.PrepareRecoveryResponseProto);</code> 17936 */ 17937 public abstract void prepareRecovery( 17938 com.google.protobuf.RpcController controller, 17939 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request, 17940 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done); 17941 17942 /** 17943 * <code>rpc acceptRecovery(.hadoop.hdfs.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.AcceptRecoveryResponseProto);</code> 17944 */ 17945 public abstract void acceptRecovery( 17946 com.google.protobuf.RpcController controller, 17947 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request, 17948 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done); 17949 17950 public static final 17951 com.google.protobuf.Descriptors.ServiceDescriptor 17952 getDescriptor() { 17953 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.getDescriptor().getServices().get(0); 17954 } 17955 public final com.google.protobuf.Descriptors.ServiceDescriptor 17956 getDescriptorForType() { 17957 return getDescriptor(); 17958 } 17959 17960 public final void callMethod( 17961 com.google.protobuf.Descriptors.MethodDescriptor method, 17962 com.google.protobuf.RpcController controller, 17963 com.google.protobuf.Message request, 17964 com.google.protobuf.RpcCallback< 17965 com.google.protobuf.Message> done) { 17966 if (method.getService() != getDescriptor()) { 17967 throw new java.lang.IllegalArgumentException( 17968 "Service.callMethod() given method descriptor for wrong " + 17969 "service type."); 17970 } 17971 switch(method.getIndex()) { 17972 case 0: 17973 this.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request, 17974 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto>specializeCallback( 17975 done)); 17976 return; 17977 case 1: 17978 this.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request, 17979 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto>specializeCallback( 17980 done)); 17981 return; 17982 case 2: 17983 this.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request, 17984 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto>specializeCallback( 17985 done)); 17986 return; 17987 case 3: 17988 this.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request, 17989 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto>specializeCallback( 17990 done)); 17991 return; 17992 case 4: 17993 this.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request, 17994 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto>specializeCallback( 17995 done)); 17996 return; 17997 case 5: 17998 this.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request, 17999 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto>specializeCallback( 18000 done)); 18001 return; 18002 case 6: 18003 this.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request, 18004 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto>specializeCallback( 18005 done)); 18006 return; 18007 case 7: 18008 this.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request, 18009 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto>specializeCallback( 18010 done)); 18011 return; 18012 case 8: 18013 this.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request, 18014 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto>specializeCallback( 18015 done)); 18016 return; 18017 case 9: 18018 this.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request, 18019 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto>specializeCallback( 18020 done)); 18021 return; 18022 case 10: 18023 this.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request, 18024 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto>specializeCallback( 18025 done)); 18026 return; 18027 case 11: 18028 this.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request, 18029 com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto>specializeCallback( 18030 done)); 18031 return; 18032 default: 18033 throw new java.lang.AssertionError("Can't get here."); 18034 } 18035 } 18036 18037 public final com.google.protobuf.Message 18038 getRequestPrototype( 18039 com.google.protobuf.Descriptors.MethodDescriptor method) { 18040 if (method.getService() != getDescriptor()) { 18041 throw new java.lang.IllegalArgumentException( 18042 "Service.getRequestPrototype() given method " + 18043 "descriptor for wrong service type."); 18044 } 18045 switch(method.getIndex()) { 18046 case 0: 18047 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance(); 18048 case 1: 18049 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance(); 18050 case 2: 18051 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance(); 18052 case 3: 18053 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance(); 18054 case 4: 18055 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance(); 18056 case 5: 18057 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); 18058 case 6: 18059 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance(); 18060 case 7: 18061 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance(); 18062 case 8: 18063 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance(); 18064 case 9: 18065 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance(); 18066 case 10: 18067 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance(); 18068 case 11: 18069 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance(); 18070 default: 18071 throw new java.lang.AssertionError("Can't get here."); 18072 } 18073 } 18074 18075 public final com.google.protobuf.Message 18076 getResponsePrototype( 18077 com.google.protobuf.Descriptors.MethodDescriptor method) { 18078 if (method.getService() != getDescriptor()) { 18079 throw new java.lang.IllegalArgumentException( 18080 "Service.getResponsePrototype() given method " + 18081 "descriptor for wrong service type."); 18082 } 18083 switch(method.getIndex()) { 18084 case 0: 18085 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(); 18086 case 1: 18087 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(); 18088 case 2: 18089 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(); 18090 case 3: 18091 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(); 18092 case 4: 18093 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(); 18094 case 5: 18095 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); 18096 case 6: 18097 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(); 18098 case 7: 18099 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(); 18100 case 8: 18101 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(); 18102 case 9: 18103 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(); 18104 case 10: 18105 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(); 18106 case 11: 18107 return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(); 18108 default: 18109 throw new java.lang.AssertionError("Can't get here."); 18110 } 18111 } 18112 18113 public static Stub newStub( 18114 com.google.protobuf.RpcChannel channel) { 18115 return new Stub(channel); 18116 } 18117 18118 public static final class Stub extends org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.QJournalProtocolService implements Interface { 18119 private Stub(com.google.protobuf.RpcChannel channel) { 18120 this.channel = channel; 18121 } 18122 18123 private final com.google.protobuf.RpcChannel channel; 18124 18125 public com.google.protobuf.RpcChannel getChannel() { 18126 return channel; 18127 } 18128 18129 public void isFormatted( 18130 com.google.protobuf.RpcController controller, 18131 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request, 18132 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) { 18133 channel.callMethod( 18134 getDescriptor().getMethods().get(0), 18135 controller, 18136 request, 18137 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(), 18138 com.google.protobuf.RpcUtil.generalizeCallback( 18139 done, 18140 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, 18141 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance())); 18142 } 18143 18144 public void getJournalState( 18145 com.google.protobuf.RpcController controller, 18146 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request, 18147 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) { 18148 channel.callMethod( 18149 getDescriptor().getMethods().get(1), 18150 controller, 18151 request, 18152 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(), 18153 com.google.protobuf.RpcUtil.generalizeCallback( 18154 done, 18155 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, 18156 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance())); 18157 } 18158 18159 public void newEpoch( 18160 com.google.protobuf.RpcController controller, 18161 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request, 18162 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) { 18163 channel.callMethod( 18164 getDescriptor().getMethods().get(2), 18165 controller, 18166 request, 18167 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(), 18168 com.google.protobuf.RpcUtil.generalizeCallback( 18169 done, 18170 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, 18171 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance())); 18172 } 18173 18174 public void format( 18175 com.google.protobuf.RpcController controller, 18176 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request, 18177 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) { 18178 channel.callMethod( 18179 getDescriptor().getMethods().get(3), 18180 controller, 18181 request, 18182 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(), 18183 com.google.protobuf.RpcUtil.generalizeCallback( 18184 done, 18185 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, 18186 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance())); 18187 } 18188 18189 public void journal( 18190 com.google.protobuf.RpcController controller, 18191 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request, 18192 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) { 18193 channel.callMethod( 18194 getDescriptor().getMethods().get(4), 18195 controller, 18196 request, 18197 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(), 18198 com.google.protobuf.RpcUtil.generalizeCallback( 18199 done, 18200 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, 18201 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance())); 18202 } 18203 18204 public void heartbeat( 18205 com.google.protobuf.RpcController controller, 18206 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request, 18207 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) { 18208 channel.callMethod( 18209 getDescriptor().getMethods().get(5), 18210 controller, 18211 request, 18212 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(), 18213 com.google.protobuf.RpcUtil.generalizeCallback( 18214 done, 18215 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, 18216 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance())); 18217 } 18218 18219 public void startLogSegment( 18220 com.google.protobuf.RpcController controller, 18221 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request, 18222 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) { 18223 channel.callMethod( 18224 getDescriptor().getMethods().get(6), 18225 controller, 18226 request, 18227 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(), 18228 com.google.protobuf.RpcUtil.generalizeCallback( 18229 done, 18230 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, 18231 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance())); 18232 } 18233 18234 public void finalizeLogSegment( 18235 com.google.protobuf.RpcController controller, 18236 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request, 18237 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) { 18238 channel.callMethod( 18239 getDescriptor().getMethods().get(7), 18240 controller, 18241 request, 18242 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(), 18243 com.google.protobuf.RpcUtil.generalizeCallback( 18244 done, 18245 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, 18246 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance())); 18247 } 18248 18249 public void purgeLogs( 18250 com.google.protobuf.RpcController controller, 18251 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request, 18252 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) { 18253 channel.callMethod( 18254 getDescriptor().getMethods().get(8), 18255 controller, 18256 request, 18257 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(), 18258 com.google.protobuf.RpcUtil.generalizeCallback( 18259 done, 18260 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, 18261 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance())); 18262 } 18263 18264 public void getEditLogManifest( 18265 com.google.protobuf.RpcController controller, 18266 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request, 18267 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) { 18268 channel.callMethod( 18269 getDescriptor().getMethods().get(9), 18270 controller, 18271 request, 18272 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(), 18273 com.google.protobuf.RpcUtil.generalizeCallback( 18274 done, 18275 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, 18276 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance())); 18277 } 18278 18279 public void prepareRecovery( 18280 com.google.protobuf.RpcController controller, 18281 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request, 18282 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) { 18283 channel.callMethod( 18284 getDescriptor().getMethods().get(10), 18285 controller, 18286 request, 18287 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(), 18288 com.google.protobuf.RpcUtil.generalizeCallback( 18289 done, 18290 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, 18291 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance())); 18292 } 18293 18294 public void acceptRecovery( 18295 com.google.protobuf.RpcController controller, 18296 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request, 18297 com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) { 18298 channel.callMethod( 18299 getDescriptor().getMethods().get(11), 18300 controller, 18301 request, 18302 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(), 18303 com.google.protobuf.RpcUtil.generalizeCallback( 18304 done, 18305 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, 18306 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance())); 18307 } 18308 } 18309 18310 public static BlockingInterface newBlockingStub( 18311 com.google.protobuf.BlockingRpcChannel channel) { 18312 return new BlockingStub(channel); 18313 } 18314 18315 public interface BlockingInterface { 18316 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted( 18317 com.google.protobuf.RpcController controller, 18318 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request) 18319 throws com.google.protobuf.ServiceException; 18320 18321 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState( 18322 com.google.protobuf.RpcController controller, 18323 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request) 18324 throws com.google.protobuf.ServiceException; 18325 18326 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch( 18327 com.google.protobuf.RpcController controller, 18328 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request) 18329 throws com.google.protobuf.ServiceException; 18330 18331 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format( 18332 com.google.protobuf.RpcController controller, 18333 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request) 18334 throws com.google.protobuf.ServiceException; 18335 18336 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal( 18337 com.google.protobuf.RpcController controller, 18338 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request) 18339 throws com.google.protobuf.ServiceException; 18340 18341 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat( 18342 com.google.protobuf.RpcController controller, 18343 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request) 18344 throws com.google.protobuf.ServiceException; 18345 18346 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment( 18347 com.google.protobuf.RpcController controller, 18348 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request) 18349 throws com.google.protobuf.ServiceException; 18350 18351 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment( 18352 com.google.protobuf.RpcController controller, 18353 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request) 18354 throws com.google.protobuf.ServiceException; 18355 18356 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs( 18357 com.google.protobuf.RpcController controller, 18358 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request) 18359 throws com.google.protobuf.ServiceException; 18360 18361 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest( 18362 com.google.protobuf.RpcController controller, 18363 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request) 18364 throws com.google.protobuf.ServiceException; 18365 18366 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery( 18367 com.google.protobuf.RpcController controller, 18368 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request) 18369 throws com.google.protobuf.ServiceException; 18370 18371 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery( 18372 com.google.protobuf.RpcController controller, 18373 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request) 18374 throws com.google.protobuf.ServiceException; 18375 } 18376 18377 private static final class BlockingStub implements BlockingInterface { 18378 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { 18379 this.channel = channel; 18380 } 18381 18382 private final com.google.protobuf.BlockingRpcChannel channel; 18383 18384 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted( 18385 com.google.protobuf.RpcController controller, 18386 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request) 18387 throws com.google.protobuf.ServiceException { 18388 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) channel.callBlockingMethod( 18389 getDescriptor().getMethods().get(0), 18390 controller, 18391 request, 18392 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()); 18393 } 18394 18395 18396 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState( 18397 com.google.protobuf.RpcController controller, 18398 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request) 18399 throws com.google.protobuf.ServiceException { 18400 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) channel.callBlockingMethod( 18401 getDescriptor().getMethods().get(1), 18402 controller, 18403 request, 18404 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()); 18405 } 18406 18407 18408 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch( 18409 com.google.protobuf.RpcController controller, 18410 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request) 18411 throws com.google.protobuf.ServiceException { 18412 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) channel.callBlockingMethod( 18413 getDescriptor().getMethods().get(2), 18414 controller, 18415 request, 18416 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()); 18417 } 18418 18419 18420 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format( 18421 com.google.protobuf.RpcController controller, 18422 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request) 18423 throws com.google.protobuf.ServiceException { 18424 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) channel.callBlockingMethod( 18425 getDescriptor().getMethods().get(3), 18426 controller, 18427 request, 18428 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()); 18429 } 18430 18431 18432 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal( 18433 com.google.protobuf.RpcController controller, 18434 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request) 18435 throws com.google.protobuf.ServiceException { 18436 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) channel.callBlockingMethod( 18437 getDescriptor().getMethods().get(4), 18438 controller, 18439 request, 18440 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()); 18441 } 18442 18443 18444 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat( 18445 com.google.protobuf.RpcController controller, 18446 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request) 18447 throws com.google.protobuf.ServiceException { 18448 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) channel.callBlockingMethod( 18449 getDescriptor().getMethods().get(5), 18450 controller, 18451 request, 18452 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()); 18453 } 18454 18455 18456 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment( 18457 com.google.protobuf.RpcController controller, 18458 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request) 18459 throws com.google.protobuf.ServiceException { 18460 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) channel.callBlockingMethod( 18461 getDescriptor().getMethods().get(6), 18462 controller, 18463 request, 18464 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()); 18465 } 18466 18467 18468 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment( 18469 com.google.protobuf.RpcController controller, 18470 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request) 18471 throws com.google.protobuf.ServiceException { 18472 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) channel.callBlockingMethod( 18473 getDescriptor().getMethods().get(7), 18474 controller, 18475 request, 18476 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()); 18477 } 18478 18479 18480 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs( 18481 com.google.protobuf.RpcController controller, 18482 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request) 18483 throws com.google.protobuf.ServiceException { 18484 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) channel.callBlockingMethod( 18485 getDescriptor().getMethods().get(8), 18486 controller, 18487 request, 18488 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()); 18489 } 18490 18491 18492 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest( 18493 com.google.protobuf.RpcController controller, 18494 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request) 18495 throws com.google.protobuf.ServiceException { 18496 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) channel.callBlockingMethod( 18497 getDescriptor().getMethods().get(9), 18498 controller, 18499 request, 18500 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()); 18501 } 18502 18503 18504 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery( 18505 com.google.protobuf.RpcController controller, 18506 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request) 18507 throws com.google.protobuf.ServiceException { 18508 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) channel.callBlockingMethod( 18509 getDescriptor().getMethods().get(10), 18510 controller, 18511 request, 18512 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()); 18513 } 18514 18515 18516 public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery( 18517 com.google.protobuf.RpcController controller, 18518 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request) 18519 throws com.google.protobuf.ServiceException { 18520 return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) channel.callBlockingMethod( 18521 getDescriptor().getMethods().get(11), 18522 controller, 18523 request, 18524 org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()); 18525 } 18526 18527 } 18528 18529 // @@protoc_insertion_point(class_scope:hadoop.hdfs.QJournalProtocolService) 18530 } 18531 18532 private static com.google.protobuf.Descriptors.Descriptor 18533 internal_static_hadoop_hdfs_JournalIdProto_descriptor; 18534 private static 18535 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18536 internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable; 18537 private static com.google.protobuf.Descriptors.Descriptor 18538 internal_static_hadoop_hdfs_RequestInfoProto_descriptor; 18539 private static 18540 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18541 internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable; 18542 private static com.google.protobuf.Descriptors.Descriptor 18543 internal_static_hadoop_hdfs_SegmentStateProto_descriptor; 18544 private static 18545 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18546 internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable; 18547 private static com.google.protobuf.Descriptors.Descriptor 18548 internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor; 18549 private static 18550 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18551 internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable; 18552 private static com.google.protobuf.Descriptors.Descriptor 18553 internal_static_hadoop_hdfs_JournalRequestProto_descriptor; 18554 private static 18555 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18556 internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable; 18557 private static com.google.protobuf.Descriptors.Descriptor 18558 internal_static_hadoop_hdfs_JournalResponseProto_descriptor; 18559 private static 18560 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18561 internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable; 18562 private static com.google.protobuf.Descriptors.Descriptor 18563 internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor; 18564 private static 18565 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18566 internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable; 18567 private static com.google.protobuf.Descriptors.Descriptor 18568 internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor; 18569 private static 18570 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18571 internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable; 18572 private static com.google.protobuf.Descriptors.Descriptor 18573 internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor; 18574 private static 18575 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18576 internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable; 18577 private static com.google.protobuf.Descriptors.Descriptor 18578 internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor; 18579 private static 18580 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18581 internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable; 18582 private static com.google.protobuf.Descriptors.Descriptor 18583 internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor; 18584 private static 18585 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18586 internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable; 18587 private static com.google.protobuf.Descriptors.Descriptor 18588 internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor; 18589 private static 18590 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18591 internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable; 18592 private static com.google.protobuf.Descriptors.Descriptor 18593 internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor; 18594 private static 18595 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18596 internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable; 18597 private static com.google.protobuf.Descriptors.Descriptor 18598 internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor; 18599 private static 18600 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18601 internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable; 18602 private static com.google.protobuf.Descriptors.Descriptor 18603 internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor; 18604 private static 18605 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18606 internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable; 18607 private static com.google.protobuf.Descriptors.Descriptor 18608 internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor; 18609 private static 18610 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18611 internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable; 18612 private static com.google.protobuf.Descriptors.Descriptor 18613 internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor; 18614 private static 18615 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18616 internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable; 18617 private static com.google.protobuf.Descriptors.Descriptor 18618 internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor; 18619 private static 18620 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18621 internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable; 18622 private static com.google.protobuf.Descriptors.Descriptor 18623 internal_static_hadoop_hdfs_FormatRequestProto_descriptor; 18624 private static 18625 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18626 internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable; 18627 private static com.google.protobuf.Descriptors.Descriptor 18628 internal_static_hadoop_hdfs_FormatResponseProto_descriptor; 18629 private static 18630 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18631 internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable; 18632 private static com.google.protobuf.Descriptors.Descriptor 18633 internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor; 18634 private static 18635 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18636 internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable; 18637 private static com.google.protobuf.Descriptors.Descriptor 18638 internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor; 18639 private static 18640 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18641 internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable; 18642 private static com.google.protobuf.Descriptors.Descriptor 18643 internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor; 18644 private static 18645 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18646 internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable; 18647 private static com.google.protobuf.Descriptors.Descriptor 18648 internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor; 18649 private static 18650 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18651 internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable; 18652 private static com.google.protobuf.Descriptors.Descriptor 18653 internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor; 18654 private static 18655 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18656 internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable; 18657 private static com.google.protobuf.Descriptors.Descriptor 18658 internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor; 18659 private static 18660 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18661 internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable; 18662 private static com.google.protobuf.Descriptors.Descriptor 18663 internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor; 18664 private static 18665 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18666 internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable; 18667 private static com.google.protobuf.Descriptors.Descriptor 18668 internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor; 18669 private static 18670 com.google.protobuf.GeneratedMessage.FieldAccessorTable 18671 internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable; 18672 18673 public static com.google.protobuf.Descriptors.FileDescriptor 18674 getDescriptor() { 18675 return descriptor; 18676 } 18677 private static com.google.protobuf.Descriptors.FileDescriptor 18678 descriptor; 18679 static { 18680 java.lang.String[] descriptorData = { 18681 "\n\026QJournalProtocol.proto\022\013hadoop.hdfs\032\nh" + 18682 "dfs.proto\"$\n\016JournalIdProto\022\022\n\nidentifie" + 18683 "r\030\001 \002(\t\"\201\001\n\020RequestInfoProto\022.\n\tjournalI" + 18684 "d\030\001 \002(\0132\033.hadoop.hdfs.JournalIdProto\022\r\n\005" + 18685 "epoch\030\002 \002(\004\022\027\n\017ipcSerialNumber\030\003 \002(\004\022\025\n\r" + 18686 "committedTxId\030\004 \001(\004\"M\n\021SegmentStateProto" + 18687 "\022\021\n\tstartTxId\030\001 \002(\004\022\017\n\007endTxId\030\002 \002(\004\022\024\n\014" + 18688 "isInProgress\030\003 \002(\010\"k\n\032PersistedRecoveryP" + 18689 "axosData\0224\n\014segmentState\030\001 \002(\0132\036.hadoop." + 18690 "hdfs.SegmentStateProto\022\027\n\017acceptedInEpoc", 18691 "h\030\002 \002(\004\"\221\001\n\023JournalRequestProto\022.\n\007reqIn" + 18692 "fo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022" + 18693 "\022\n\nfirstTxnId\030\002 \002(\004\022\017\n\007numTxns\030\003 \002(\r\022\017\n\007" + 18694 "records\030\004 \002(\014\022\024\n\014segmentTxnId\030\005 \002(\004\"\026\n\024J" + 18695 "ournalResponseProto\"G\n\025HeartbeatRequestP" + 18696 "roto\022.\n\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.Requ" + 18697 "estInfoProto\"\030\n\026HeartbeatResponseProto\"[" + 18698 "\n\033StartLogSegmentRequestProto\022.\n\007reqInfo" + 18699 "\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022\014\n" + 18700 "\004txid\030\002 \002(\004\"\036\n\034StartLogSegmentResponsePr", 18701 "oto\"t\n\036FinalizeLogSegmentRequestProto\022.\n" + 18702 "\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfo" + 18703 "Proto\022\021\n\tstartTxId\030\002 \002(\004\022\017\n\007endTxId\030\003 \002(" + 18704 "\004\"!\n\037FinalizeLogSegmentResponseProto\"^\n\025" + 18705 "PurgeLogsRequestProto\022.\n\007reqInfo\030\001 \002(\0132\035" + 18706 ".hadoop.hdfs.RequestInfoProto\022\025\n\rminTxId" + 18707 "ToKeep\030\002 \002(\004\"\030\n\026PurgeLogsResponseProto\"C" + 18708 "\n\027IsFormattedRequestProto\022(\n\003jid\030\001 \002(\0132\033" + 18709 ".hadoop.hdfs.JournalIdProto\"/\n\030IsFormatt" + 18710 "edResponseProto\022\023\n\013isFormatted\030\001 \002(\010\"G\n\033", 18711 "GetJournalStateRequestProto\022(\n\003jid\030\001 \002(\013" + 18712 "2\033.hadoop.hdfs.JournalIdProto\"\\\n\034GetJour" + 18713 "nalStateResponseProto\022\031\n\021lastPromisedEpo" + 18714 "ch\030\001 \002(\004\022\020\n\010httpPort\030\002 \002(\r\022\017\n\007fromURL\030\003 " + 18715 "\001(\t\"o\n\022FormatRequestProto\022(\n\003jid\030\001 \002(\0132\033" + 18716 ".hadoop.hdfs.JournalIdProto\022/\n\006nsInfo\030\002 " + 18717 "\002(\0132\037.hadoop.hdfs.NamespaceInfoProto\"\025\n\023" + 18718 "FormatResponseProto\"\200\001\n\024NewEpochRequestP" + 18719 "roto\022(\n\003jid\030\001 \002(\0132\033.hadoop.hdfs.JournalI" + 18720 "dProto\022/\n\006nsInfo\030\002 \002(\0132\037.hadoop.hdfs.Nam", 18721 "espaceInfoProto\022\r\n\005epoch\030\003 \002(\004\"0\n\025NewEpo" + 18722 "chResponseProto\022\027\n\017lastSegmentTxId\030\001 \001(\004" + 18723 "\"z\n\036GetEditLogManifestRequestProto\022(\n\003ji" + 18724 "d\030\001 \002(\0132\033.hadoop.hdfs.JournalIdProto\022\021\n\t" + 18725 "sinceTxId\030\002 \002(\004\022\033\n\014inProgressOk\030\004 \001(\010:\005f" + 18726 "alse\"\177\n\037GetEditLogManifestResponseProto\022" + 18727 "9\n\010manifest\030\001 \002(\0132\'.hadoop.hdfs.RemoteEd" + 18728 "itLogManifestProto\022\020\n\010httpPort\030\002 \002(\r\022\017\n\007" + 18729 "fromURL\030\003 \001(\t\"b\n\033PrepareRecoveryRequestP" + 18730 "roto\022.\n\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.Requ", 18731 "estInfoProto\022\023\n\013segmentTxId\030\002 \002(\004\"\241\001\n\034Pr" + 18732 "epareRecoveryResponseProto\0224\n\014segmentSta" + 18733 "te\030\001 \001(\0132\036.hadoop.hdfs.SegmentStateProto" + 18734 "\022\027\n\017acceptedInEpoch\030\002 \001(\004\022\027\n\017lastWriterE" + 18735 "poch\030\003 \002(\004\022\031\n\021lastCommittedTxId\030\004 \001(\004\"\224\001" + 18736 "\n\032AcceptRecoveryRequestProto\022.\n\007reqInfo\030" + 18737 "\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\0225\n\r" + 18738 "stateToAccept\030\002 \002(\0132\036.hadoop.hdfs.Segmen" + 18739 "tStateProto\022\017\n\007fromURL\030\003 \002(\t\"\035\n\033AcceptRe" + 18740 "coveryResponseProto2\220\t\n\027QJournalProtocol", 18741 "Service\022Z\n\013isFormatted\022$.hadoop.hdfs.IsF" + 18742 "ormattedRequestProto\032%.hadoop.hdfs.IsFor" + 18743 "mattedResponseProto\022f\n\017getJournalState\022(" + 18744 ".hadoop.hdfs.GetJournalStateRequestProto" + 18745 "\032).hadoop.hdfs.GetJournalStateResponsePr" + 18746 "oto\022Q\n\010newEpoch\022!.hadoop.hdfs.NewEpochRe" + 18747 "questProto\032\".hadoop.hdfs.NewEpochRespons" + 18748 "eProto\022K\n\006format\022\037.hadoop.hdfs.FormatReq" + 18749 "uestProto\032 .hadoop.hdfs.FormatResponsePr" + 18750 "oto\022N\n\007journal\022 .hadoop.hdfs.JournalRequ", 18751 "estProto\032!.hadoop.hdfs.JournalResponsePr" + 18752 "oto\022T\n\theartbeat\022\".hadoop.hdfs.Heartbeat" + 18753 "RequestProto\032#.hadoop.hdfs.HeartbeatResp" + 18754 "onseProto\022f\n\017startLogSegment\022(.hadoop.hd" + 18755 "fs.StartLogSegmentRequestProto\032).hadoop." + 18756 "hdfs.StartLogSegmentResponseProto\022o\n\022fin" + 18757 "alizeLogSegment\022+.hadoop.hdfs.FinalizeLo" + 18758 "gSegmentRequestProto\032,.hadoop.hdfs.Final" + 18759 "izeLogSegmentResponseProto\022T\n\tpurgeLogs\022" + 18760 "\".hadoop.hdfs.PurgeLogsRequestProto\032#.ha", 18761 "doop.hdfs.PurgeLogsResponseProto\022o\n\022getE" + 18762 "ditLogManifest\022+.hadoop.hdfs.GetEditLogM" + 18763 "anifestRequestProto\032,.hadoop.hdfs.GetEdi" + 18764 "tLogManifestResponseProto\022f\n\017prepareReco" + 18765 "very\022(.hadoop.hdfs.PrepareRecoveryReques" + 18766 "tProto\032).hadoop.hdfs.PrepareRecoveryResp" + 18767 "onseProto\022c\n\016acceptRecovery\022\'.hadoop.hdf" + 18768 "s.AcceptRecoveryRequestProto\032(.hadoop.hd" + 18769 "fs.AcceptRecoveryResponseProtoBH\n(org.ap" + 18770 "ache.hadoop.hdfs.qjournal.protocolB\026QJou", 18771 "rnalProtocolProtos\210\001\001\240\001\001" 18772 }; 18773 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 18774 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 18775 public com.google.protobuf.ExtensionRegistry assignDescriptors( 18776 com.google.protobuf.Descriptors.FileDescriptor root) { 18777 descriptor = root; 18778 internal_static_hadoop_hdfs_JournalIdProto_descriptor = 18779 getDescriptor().getMessageTypes().get(0); 18780 internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable = new 18781 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18782 internal_static_hadoop_hdfs_JournalIdProto_descriptor, 18783 new java.lang.String[] { "Identifier", }); 18784 internal_static_hadoop_hdfs_RequestInfoProto_descriptor = 18785 getDescriptor().getMessageTypes().get(1); 18786 internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable = new 18787 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18788 internal_static_hadoop_hdfs_RequestInfoProto_descriptor, 18789 new java.lang.String[] { "JournalId", "Epoch", "IpcSerialNumber", "CommittedTxId", }); 18790 internal_static_hadoop_hdfs_SegmentStateProto_descriptor = 18791 getDescriptor().getMessageTypes().get(2); 18792 internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable = new 18793 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18794 internal_static_hadoop_hdfs_SegmentStateProto_descriptor, 18795 new java.lang.String[] { "StartTxId", "EndTxId", "IsInProgress", }); 18796 internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor = 18797 getDescriptor().getMessageTypes().get(3); 18798 internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable = new 18799 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18800 internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor, 18801 new java.lang.String[] { "SegmentState", "AcceptedInEpoch", }); 18802 internal_static_hadoop_hdfs_JournalRequestProto_descriptor = 18803 getDescriptor().getMessageTypes().get(4); 18804 internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable = new 18805 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18806 internal_static_hadoop_hdfs_JournalRequestProto_descriptor, 18807 new java.lang.String[] { "ReqInfo", "FirstTxnId", "NumTxns", "Records", "SegmentTxnId", }); 18808 internal_static_hadoop_hdfs_JournalResponseProto_descriptor = 18809 getDescriptor().getMessageTypes().get(5); 18810 internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable = new 18811 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18812 internal_static_hadoop_hdfs_JournalResponseProto_descriptor, 18813 new java.lang.String[] { }); 18814 internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor = 18815 getDescriptor().getMessageTypes().get(6); 18816 internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable = new 18817 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18818 internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor, 18819 new java.lang.String[] { "ReqInfo", }); 18820 internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor = 18821 getDescriptor().getMessageTypes().get(7); 18822 internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable = new 18823 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18824 internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor, 18825 new java.lang.String[] { }); 18826 internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor = 18827 getDescriptor().getMessageTypes().get(8); 18828 internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable = new 18829 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18830 internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor, 18831 new java.lang.String[] { "ReqInfo", "Txid", }); 18832 internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor = 18833 getDescriptor().getMessageTypes().get(9); 18834 internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable = new 18835 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18836 internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor, 18837 new java.lang.String[] { }); 18838 internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor = 18839 getDescriptor().getMessageTypes().get(10); 18840 internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable = new 18841 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18842 internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor, 18843 new java.lang.String[] { "ReqInfo", "StartTxId", "EndTxId", }); 18844 internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor = 18845 getDescriptor().getMessageTypes().get(11); 18846 internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable = new 18847 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18848 internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor, 18849 new java.lang.String[] { }); 18850 internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor = 18851 getDescriptor().getMessageTypes().get(12); 18852 internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable = new 18853 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18854 internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor, 18855 new java.lang.String[] { "ReqInfo", "MinTxIdToKeep", }); 18856 internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor = 18857 getDescriptor().getMessageTypes().get(13); 18858 internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable = new 18859 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18860 internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor, 18861 new java.lang.String[] { }); 18862 internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor = 18863 getDescriptor().getMessageTypes().get(14); 18864 internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable = new 18865 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18866 internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor, 18867 new java.lang.String[] { "Jid", }); 18868 internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor = 18869 getDescriptor().getMessageTypes().get(15); 18870 internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable = new 18871 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18872 internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor, 18873 new java.lang.String[] { "IsFormatted", }); 18874 internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor = 18875 getDescriptor().getMessageTypes().get(16); 18876 internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable = new 18877 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18878 internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor, 18879 new java.lang.String[] { "Jid", }); 18880 internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor = 18881 getDescriptor().getMessageTypes().get(17); 18882 internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable = new 18883 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18884 internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor, 18885 new java.lang.String[] { "LastPromisedEpoch", "HttpPort", "FromURL", }); 18886 internal_static_hadoop_hdfs_FormatRequestProto_descriptor = 18887 getDescriptor().getMessageTypes().get(18); 18888 internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable = new 18889 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18890 internal_static_hadoop_hdfs_FormatRequestProto_descriptor, 18891 new java.lang.String[] { "Jid", "NsInfo", }); 18892 internal_static_hadoop_hdfs_FormatResponseProto_descriptor = 18893 getDescriptor().getMessageTypes().get(19); 18894 internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable = new 18895 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18896 internal_static_hadoop_hdfs_FormatResponseProto_descriptor, 18897 new java.lang.String[] { }); 18898 internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor = 18899 getDescriptor().getMessageTypes().get(20); 18900 internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable = new 18901 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18902 internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor, 18903 new java.lang.String[] { "Jid", "NsInfo", "Epoch", }); 18904 internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor = 18905 getDescriptor().getMessageTypes().get(21); 18906 internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable = new 18907 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18908 internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor, 18909 new java.lang.String[] { "LastSegmentTxId", }); 18910 internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor = 18911 getDescriptor().getMessageTypes().get(22); 18912 internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable = new 18913 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18914 internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor, 18915 new java.lang.String[] { "Jid", "SinceTxId", "InProgressOk", }); 18916 internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor = 18917 getDescriptor().getMessageTypes().get(23); 18918 internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable = new 18919 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18920 internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor, 18921 new java.lang.String[] { "Manifest", "HttpPort", "FromURL", }); 18922 internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor = 18923 getDescriptor().getMessageTypes().get(24); 18924 internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable = new 18925 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18926 internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor, 18927 new java.lang.String[] { "ReqInfo", "SegmentTxId", }); 18928 internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor = 18929 getDescriptor().getMessageTypes().get(25); 18930 internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable = new 18931 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18932 internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor, 18933 new java.lang.String[] { "SegmentState", "AcceptedInEpoch", "LastWriterEpoch", "LastCommittedTxId", }); 18934 internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor = 18935 getDescriptor().getMessageTypes().get(26); 18936 internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable = new 18937 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18938 internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor, 18939 new java.lang.String[] { "ReqInfo", "StateToAccept", "FromURL", }); 18940 internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor = 18941 getDescriptor().getMessageTypes().get(27); 18942 internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable = new 18943 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 18944 internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor, 18945 new java.lang.String[] { }); 18946 return null; 18947 } 18948 }; 18949 com.google.protobuf.Descriptors.FileDescriptor 18950 .internalBuildGeneratedFileFrom(descriptorData, 18951 new com.google.protobuf.Descriptors.FileDescriptor[] { 18952 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(), 18953 }, assigner); 18954 } 18955 18956 // @@protoc_insertion_point(outer_class_scope) 18957 }