001 // Generated by the protocol buffer compiler. DO NOT EDIT! 002 // source: fsimage.proto 003 004 package org.apache.hadoop.hdfs.server.namenode; 005 006 public final class FsImageProto { 007 private FsImageProto() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public interface FileSummaryOrBuilder 012 extends com.google.protobuf.MessageOrBuilder { 013 014 // required uint32 ondiskVersion = 1; 015 /** 016 * <code>required uint32 ondiskVersion = 1;</code> 017 * 018 * <pre> 019 * The version of the above EBNF grammars. 020 * </pre> 021 */ 022 boolean hasOndiskVersion(); 023 /** 024 * <code>required uint32 ondiskVersion = 1;</code> 025 * 026 * <pre> 027 * The version of the above EBNF grammars. 028 * </pre> 029 */ 030 int getOndiskVersion(); 031 032 // required uint32 layoutVersion = 2; 033 /** 034 * <code>required uint32 layoutVersion = 2;</code> 035 * 036 * <pre> 037 * layoutVersion describes which features are available in the 038 * FSImage. 039 * </pre> 040 */ 041 boolean hasLayoutVersion(); 042 /** 043 * <code>required uint32 layoutVersion = 2;</code> 044 * 045 * <pre> 046 * layoutVersion describes which features are available in the 047 * FSImage. 048 * </pre> 049 */ 050 int getLayoutVersion(); 051 052 // optional string codec = 3; 053 /** 054 * <code>optional string codec = 3;</code> 055 */ 056 boolean hasCodec(); 057 /** 058 * <code>optional string codec = 3;</code> 059 */ 060 java.lang.String getCodec(); 061 /** 062 * <code>optional string codec = 3;</code> 063 */ 064 com.google.protobuf.ByteString 065 getCodecBytes(); 066 067 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 068 /** 069 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 070 */ 071 java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> 072 getSectionsList(); 073 /** 074 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 075 */ 076 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index); 077 /** 078 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 079 */ 080 int getSectionsCount(); 081 /** 082 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 083 */ 084 java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 085 getSectionsOrBuilderList(); 086 /** 087 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 088 */ 089 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 090 int index); 091 } 092 /** 093 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary} 094 */ 095 public static final class FileSummary extends 096 com.google.protobuf.GeneratedMessage 097 implements FileSummaryOrBuilder { 098 // Use FileSummary.newBuilder() to construct. 099 private FileSummary(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 100 super(builder); 101 this.unknownFields = builder.getUnknownFields(); 102 } 103 private FileSummary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 104 105 private static final FileSummary defaultInstance; 106 public static FileSummary getDefaultInstance() { 107 return defaultInstance; 108 } 109 110 public FileSummary getDefaultInstanceForType() { 111 return defaultInstance; 112 } 113 114 private final com.google.protobuf.UnknownFieldSet unknownFields; 115 @java.lang.Override 116 public final com.google.protobuf.UnknownFieldSet 117 getUnknownFields() { 118 return this.unknownFields; 119 } 120 private FileSummary( 121 com.google.protobuf.CodedInputStream input, 122 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 123 throws com.google.protobuf.InvalidProtocolBufferException { 124 initFields(); 125 int mutable_bitField0_ = 0; 126 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 127 com.google.protobuf.UnknownFieldSet.newBuilder(); 128 try { 129 boolean done = false; 130 while (!done) { 131 int tag = input.readTag(); 132 switch (tag) { 133 case 0: 134 done = true; 135 break; 136 default: { 137 if (!parseUnknownField(input, unknownFields, 138 extensionRegistry, tag)) { 139 done = true; 140 } 141 break; 142 } 143 case 8: { 144 bitField0_ |= 0x00000001; 145 ondiskVersion_ = input.readUInt32(); 146 break; 147 } 148 case 16: { 149 bitField0_ |= 0x00000002; 150 layoutVersion_ = input.readUInt32(); 151 break; 152 } 153 case 26: { 154 bitField0_ |= 0x00000004; 155 codec_ = input.readBytes(); 156 break; 157 } 158 case 34: { 159 if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) { 160 sections_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section>(); 161 mutable_bitField0_ |= 0x00000008; 162 } 163 sections_.add(input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.PARSER, extensionRegistry)); 164 break; 165 } 166 } 167 } 168 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 169 throw e.setUnfinishedMessage(this); 170 } catch (java.io.IOException e) { 171 throw new com.google.protobuf.InvalidProtocolBufferException( 172 e.getMessage()).setUnfinishedMessage(this); 173 } finally { 174 if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) { 175 sections_ = java.util.Collections.unmodifiableList(sections_); 176 } 177 this.unknownFields = unknownFields.build(); 178 makeExtensionsImmutable(); 179 } 180 } 181 public static final com.google.protobuf.Descriptors.Descriptor 182 getDescriptor() { 183 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 184 } 185 186 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 187 internalGetFieldAccessorTable() { 188 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable 189 .ensureFieldAccessorsInitialized( 190 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Builder.class); 191 } 192 193 public static com.google.protobuf.Parser<FileSummary> PARSER = 194 new com.google.protobuf.AbstractParser<FileSummary>() { 195 public FileSummary parsePartialFrom( 196 com.google.protobuf.CodedInputStream input, 197 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 198 throws com.google.protobuf.InvalidProtocolBufferException { 199 return new FileSummary(input, extensionRegistry); 200 } 201 }; 202 203 @java.lang.Override 204 public com.google.protobuf.Parser<FileSummary> getParserForType() { 205 return PARSER; 206 } 207 208 public interface SectionOrBuilder 209 extends com.google.protobuf.MessageOrBuilder { 210 211 // optional string name = 1; 212 /** 213 * <code>optional string name = 1;</code> 214 */ 215 boolean hasName(); 216 /** 217 * <code>optional string name = 1;</code> 218 */ 219 java.lang.String getName(); 220 /** 221 * <code>optional string name = 1;</code> 222 */ 223 com.google.protobuf.ByteString 224 getNameBytes(); 225 226 // optional uint64 length = 2; 227 /** 228 * <code>optional uint64 length = 2;</code> 229 */ 230 boolean hasLength(); 231 /** 232 * <code>optional uint64 length = 2;</code> 233 */ 234 long getLength(); 235 236 // optional uint64 offset = 3; 237 /** 238 * <code>optional uint64 offset = 3;</code> 239 */ 240 boolean hasOffset(); 241 /** 242 * <code>optional uint64 offset = 3;</code> 243 */ 244 long getOffset(); 245 } 246 /** 247 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary.Section} 248 * 249 * <pre> 250 * index for each section 251 * </pre> 252 */ 253 public static final class Section extends 254 com.google.protobuf.GeneratedMessage 255 implements SectionOrBuilder { 256 // Use Section.newBuilder() to construct. 257 private Section(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 258 super(builder); 259 this.unknownFields = builder.getUnknownFields(); 260 } 261 private Section(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 262 263 private static final Section defaultInstance; 264 public static Section getDefaultInstance() { 265 return defaultInstance; 266 } 267 268 public Section getDefaultInstanceForType() { 269 return defaultInstance; 270 } 271 272 private final com.google.protobuf.UnknownFieldSet unknownFields; 273 @java.lang.Override 274 public final com.google.protobuf.UnknownFieldSet 275 getUnknownFields() { 276 return this.unknownFields; 277 } 278 private Section( 279 com.google.protobuf.CodedInputStream input, 280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 281 throws com.google.protobuf.InvalidProtocolBufferException { 282 initFields(); 283 int mutable_bitField0_ = 0; 284 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 285 com.google.protobuf.UnknownFieldSet.newBuilder(); 286 try { 287 boolean done = false; 288 while (!done) { 289 int tag = input.readTag(); 290 switch (tag) { 291 case 0: 292 done = true; 293 break; 294 default: { 295 if (!parseUnknownField(input, unknownFields, 296 extensionRegistry, tag)) { 297 done = true; 298 } 299 break; 300 } 301 case 10: { 302 bitField0_ |= 0x00000001; 303 name_ = input.readBytes(); 304 break; 305 } 306 case 16: { 307 bitField0_ |= 0x00000002; 308 length_ = input.readUInt64(); 309 break; 310 } 311 case 24: { 312 bitField0_ |= 0x00000004; 313 offset_ = input.readUInt64(); 314 break; 315 } 316 } 317 } 318 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 319 throw e.setUnfinishedMessage(this); 320 } catch (java.io.IOException e) { 321 throw new com.google.protobuf.InvalidProtocolBufferException( 322 e.getMessage()).setUnfinishedMessage(this); 323 } finally { 324 this.unknownFields = unknownFields.build(); 325 makeExtensionsImmutable(); 326 } 327 } 328 public static final com.google.protobuf.Descriptors.Descriptor 329 getDescriptor() { 330 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 331 } 332 333 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 334 internalGetFieldAccessorTable() { 335 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable 336 .ensureFieldAccessorsInitialized( 337 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder.class); 338 } 339 340 public static com.google.protobuf.Parser<Section> PARSER = 341 new com.google.protobuf.AbstractParser<Section>() { 342 public Section parsePartialFrom( 343 com.google.protobuf.CodedInputStream input, 344 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 345 throws com.google.protobuf.InvalidProtocolBufferException { 346 return new Section(input, extensionRegistry); 347 } 348 }; 349 350 @java.lang.Override 351 public com.google.protobuf.Parser<Section> getParserForType() { 352 return PARSER; 353 } 354 355 private int bitField0_; 356 // optional string name = 1; 357 public static final int NAME_FIELD_NUMBER = 1; 358 private java.lang.Object name_; 359 /** 360 * <code>optional string name = 1;</code> 361 */ 362 public boolean hasName() { 363 return ((bitField0_ & 0x00000001) == 0x00000001); 364 } 365 /** 366 * <code>optional string name = 1;</code> 367 */ 368 public java.lang.String getName() { 369 java.lang.Object ref = name_; 370 if (ref instanceof java.lang.String) { 371 return (java.lang.String) ref; 372 } else { 373 com.google.protobuf.ByteString bs = 374 (com.google.protobuf.ByteString) ref; 375 java.lang.String s = bs.toStringUtf8(); 376 if (bs.isValidUtf8()) { 377 name_ = s; 378 } 379 return s; 380 } 381 } 382 /** 383 * <code>optional string name = 1;</code> 384 */ 385 public com.google.protobuf.ByteString 386 getNameBytes() { 387 java.lang.Object ref = name_; 388 if (ref instanceof java.lang.String) { 389 com.google.protobuf.ByteString b = 390 com.google.protobuf.ByteString.copyFromUtf8( 391 (java.lang.String) ref); 392 name_ = b; 393 return b; 394 } else { 395 return (com.google.protobuf.ByteString) ref; 396 } 397 } 398 399 // optional uint64 length = 2; 400 public static final int LENGTH_FIELD_NUMBER = 2; 401 private long length_; 402 /** 403 * <code>optional uint64 length = 2;</code> 404 */ 405 public boolean hasLength() { 406 return ((bitField0_ & 0x00000002) == 0x00000002); 407 } 408 /** 409 * <code>optional uint64 length = 2;</code> 410 */ 411 public long getLength() { 412 return length_; 413 } 414 415 // optional uint64 offset = 3; 416 public static final int OFFSET_FIELD_NUMBER = 3; 417 private long offset_; 418 /** 419 * <code>optional uint64 offset = 3;</code> 420 */ 421 public boolean hasOffset() { 422 return ((bitField0_ & 0x00000004) == 0x00000004); 423 } 424 /** 425 * <code>optional uint64 offset = 3;</code> 426 */ 427 public long getOffset() { 428 return offset_; 429 } 430 431 private void initFields() { 432 name_ = ""; 433 length_ = 0L; 434 offset_ = 0L; 435 } 436 private byte memoizedIsInitialized = -1; 437 public final boolean isInitialized() { 438 byte isInitialized = memoizedIsInitialized; 439 if (isInitialized != -1) return isInitialized == 1; 440 441 memoizedIsInitialized = 1; 442 return true; 443 } 444 445 public void writeTo(com.google.protobuf.CodedOutputStream output) 446 throws java.io.IOException { 447 getSerializedSize(); 448 if (((bitField0_ & 0x00000001) == 0x00000001)) { 449 output.writeBytes(1, getNameBytes()); 450 } 451 if (((bitField0_ & 0x00000002) == 0x00000002)) { 452 output.writeUInt64(2, length_); 453 } 454 if (((bitField0_ & 0x00000004) == 0x00000004)) { 455 output.writeUInt64(3, offset_); 456 } 457 getUnknownFields().writeTo(output); 458 } 459 460 private int memoizedSerializedSize = -1; 461 public int getSerializedSize() { 462 int size = memoizedSerializedSize; 463 if (size != -1) return size; 464 465 size = 0; 466 if (((bitField0_ & 0x00000001) == 0x00000001)) { 467 size += com.google.protobuf.CodedOutputStream 468 .computeBytesSize(1, getNameBytes()); 469 } 470 if (((bitField0_ & 0x00000002) == 0x00000002)) { 471 size += com.google.protobuf.CodedOutputStream 472 .computeUInt64Size(2, length_); 473 } 474 if (((bitField0_ & 0x00000004) == 0x00000004)) { 475 size += com.google.protobuf.CodedOutputStream 476 .computeUInt64Size(3, offset_); 477 } 478 size += getUnknownFields().getSerializedSize(); 479 memoizedSerializedSize = size; 480 return size; 481 } 482 483 private static final long serialVersionUID = 0L; 484 @java.lang.Override 485 protected java.lang.Object writeReplace() 486 throws java.io.ObjectStreamException { 487 return super.writeReplace(); 488 } 489 490 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 491 com.google.protobuf.ByteString data) 492 throws com.google.protobuf.InvalidProtocolBufferException { 493 return PARSER.parseFrom(data); 494 } 495 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 496 com.google.protobuf.ByteString data, 497 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 498 throws com.google.protobuf.InvalidProtocolBufferException { 499 return PARSER.parseFrom(data, extensionRegistry); 500 } 501 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom(byte[] data) 502 throws com.google.protobuf.InvalidProtocolBufferException { 503 return PARSER.parseFrom(data); 504 } 505 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 506 byte[] data, 507 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 508 throws com.google.protobuf.InvalidProtocolBufferException { 509 return PARSER.parseFrom(data, extensionRegistry); 510 } 511 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom(java.io.InputStream input) 512 throws java.io.IOException { 513 return PARSER.parseFrom(input); 514 } 515 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 516 java.io.InputStream input, 517 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 518 throws java.io.IOException { 519 return PARSER.parseFrom(input, extensionRegistry); 520 } 521 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseDelimitedFrom(java.io.InputStream input) 522 throws java.io.IOException { 523 return PARSER.parseDelimitedFrom(input); 524 } 525 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseDelimitedFrom( 526 java.io.InputStream input, 527 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 528 throws java.io.IOException { 529 return PARSER.parseDelimitedFrom(input, extensionRegistry); 530 } 531 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 532 com.google.protobuf.CodedInputStream input) 533 throws java.io.IOException { 534 return PARSER.parseFrom(input); 535 } 536 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 537 com.google.protobuf.CodedInputStream input, 538 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 539 throws java.io.IOException { 540 return PARSER.parseFrom(input, extensionRegistry); 541 } 542 543 public static Builder newBuilder() { return Builder.create(); } 544 public Builder newBuilderForType() { return newBuilder(); } 545 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section prototype) { 546 return newBuilder().mergeFrom(prototype); 547 } 548 public Builder toBuilder() { return newBuilder(this); } 549 550 @java.lang.Override 551 protected Builder newBuilderForType( 552 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 553 Builder builder = new Builder(parent); 554 return builder; 555 } 556 /** 557 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary.Section} 558 * 559 * <pre> 560 * index for each section 561 * </pre> 562 */ 563 public static final class Builder extends 564 com.google.protobuf.GeneratedMessage.Builder<Builder> 565 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder { 566 public static final com.google.protobuf.Descriptors.Descriptor 567 getDescriptor() { 568 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 569 } 570 571 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 572 internalGetFieldAccessorTable() { 573 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable 574 .ensureFieldAccessorsInitialized( 575 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder.class); 576 } 577 578 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.newBuilder() 579 private Builder() { 580 maybeForceBuilderInitialization(); 581 } 582 583 private Builder( 584 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 585 super(parent); 586 maybeForceBuilderInitialization(); 587 } 588 private void maybeForceBuilderInitialization() { 589 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 590 } 591 } 592 private static Builder create() { 593 return new Builder(); 594 } 595 596 public Builder clear() { 597 super.clear(); 598 name_ = ""; 599 bitField0_ = (bitField0_ & ~0x00000001); 600 length_ = 0L; 601 bitField0_ = (bitField0_ & ~0x00000002); 602 offset_ = 0L; 603 bitField0_ = (bitField0_ & ~0x00000004); 604 return this; 605 } 606 607 public Builder clone() { 608 return create().mergeFrom(buildPartial()); 609 } 610 611 public com.google.protobuf.Descriptors.Descriptor 612 getDescriptorForType() { 613 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 614 } 615 616 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getDefaultInstanceForType() { 617 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance(); 618 } 619 620 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section build() { 621 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section result = buildPartial(); 622 if (!result.isInitialized()) { 623 throw newUninitializedMessageException(result); 624 } 625 return result; 626 } 627 628 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section buildPartial() { 629 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section(this); 630 int from_bitField0_ = bitField0_; 631 int to_bitField0_ = 0; 632 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 633 to_bitField0_ |= 0x00000001; 634 } 635 result.name_ = name_; 636 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 637 to_bitField0_ |= 0x00000002; 638 } 639 result.length_ = length_; 640 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 641 to_bitField0_ |= 0x00000004; 642 } 643 result.offset_ = offset_; 644 result.bitField0_ = to_bitField0_; 645 onBuilt(); 646 return result; 647 } 648 649 public Builder mergeFrom(com.google.protobuf.Message other) { 650 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section) { 651 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section)other); 652 } else { 653 super.mergeFrom(other); 654 return this; 655 } 656 } 657 658 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section other) { 659 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()) return this; 660 if (other.hasName()) { 661 bitField0_ |= 0x00000001; 662 name_ = other.name_; 663 onChanged(); 664 } 665 if (other.hasLength()) { 666 setLength(other.getLength()); 667 } 668 if (other.hasOffset()) { 669 setOffset(other.getOffset()); 670 } 671 this.mergeUnknownFields(other.getUnknownFields()); 672 return this; 673 } 674 675 public final boolean isInitialized() { 676 return true; 677 } 678 679 public Builder mergeFrom( 680 com.google.protobuf.CodedInputStream input, 681 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 682 throws java.io.IOException { 683 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parsedMessage = null; 684 try { 685 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 686 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 687 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section) e.getUnfinishedMessage(); 688 throw e; 689 } finally { 690 if (parsedMessage != null) { 691 mergeFrom(parsedMessage); 692 } 693 } 694 return this; 695 } 696 private int bitField0_; 697 698 // optional string name = 1; 699 private java.lang.Object name_ = ""; 700 /** 701 * <code>optional string name = 1;</code> 702 */ 703 public boolean hasName() { 704 return ((bitField0_ & 0x00000001) == 0x00000001); 705 } 706 /** 707 * <code>optional string name = 1;</code> 708 */ 709 public java.lang.String getName() { 710 java.lang.Object ref = name_; 711 if (!(ref instanceof java.lang.String)) { 712 java.lang.String s = ((com.google.protobuf.ByteString) ref) 713 .toStringUtf8(); 714 name_ = s; 715 return s; 716 } else { 717 return (java.lang.String) ref; 718 } 719 } 720 /** 721 * <code>optional string name = 1;</code> 722 */ 723 public com.google.protobuf.ByteString 724 getNameBytes() { 725 java.lang.Object ref = name_; 726 if (ref instanceof String) { 727 com.google.protobuf.ByteString b = 728 com.google.protobuf.ByteString.copyFromUtf8( 729 (java.lang.String) ref); 730 name_ = b; 731 return b; 732 } else { 733 return (com.google.protobuf.ByteString) ref; 734 } 735 } 736 /** 737 * <code>optional string name = 1;</code> 738 */ 739 public Builder setName( 740 java.lang.String value) { 741 if (value == null) { 742 throw new NullPointerException(); 743 } 744 bitField0_ |= 0x00000001; 745 name_ = value; 746 onChanged(); 747 return this; 748 } 749 /** 750 * <code>optional string name = 1;</code> 751 */ 752 public Builder clearName() { 753 bitField0_ = (bitField0_ & ~0x00000001); 754 name_ = getDefaultInstance().getName(); 755 onChanged(); 756 return this; 757 } 758 /** 759 * <code>optional string name = 1;</code> 760 */ 761 public Builder setNameBytes( 762 com.google.protobuf.ByteString value) { 763 if (value == null) { 764 throw new NullPointerException(); 765 } 766 bitField0_ |= 0x00000001; 767 name_ = value; 768 onChanged(); 769 return this; 770 } 771 772 // optional uint64 length = 2; 773 private long length_ ; 774 /** 775 * <code>optional uint64 length = 2;</code> 776 */ 777 public boolean hasLength() { 778 return ((bitField0_ & 0x00000002) == 0x00000002); 779 } 780 /** 781 * <code>optional uint64 length = 2;</code> 782 */ 783 public long getLength() { 784 return length_; 785 } 786 /** 787 * <code>optional uint64 length = 2;</code> 788 */ 789 public Builder setLength(long value) { 790 bitField0_ |= 0x00000002; 791 length_ = value; 792 onChanged(); 793 return this; 794 } 795 /** 796 * <code>optional uint64 length = 2;</code> 797 */ 798 public Builder clearLength() { 799 bitField0_ = (bitField0_ & ~0x00000002); 800 length_ = 0L; 801 onChanged(); 802 return this; 803 } 804 805 // optional uint64 offset = 3; 806 private long offset_ ; 807 /** 808 * <code>optional uint64 offset = 3;</code> 809 */ 810 public boolean hasOffset() { 811 return ((bitField0_ & 0x00000004) == 0x00000004); 812 } 813 /** 814 * <code>optional uint64 offset = 3;</code> 815 */ 816 public long getOffset() { 817 return offset_; 818 } 819 /** 820 * <code>optional uint64 offset = 3;</code> 821 */ 822 public Builder setOffset(long value) { 823 bitField0_ |= 0x00000004; 824 offset_ = value; 825 onChanged(); 826 return this; 827 } 828 /** 829 * <code>optional uint64 offset = 3;</code> 830 */ 831 public Builder clearOffset() { 832 bitField0_ = (bitField0_ & ~0x00000004); 833 offset_ = 0L; 834 onChanged(); 835 return this; 836 } 837 838 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FileSummary.Section) 839 } 840 841 static { 842 defaultInstance = new Section(true); 843 defaultInstance.initFields(); 844 } 845 846 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FileSummary.Section) 847 } 848 849 private int bitField0_; 850 // required uint32 ondiskVersion = 1; 851 public static final int ONDISKVERSION_FIELD_NUMBER = 1; 852 private int ondiskVersion_; 853 /** 854 * <code>required uint32 ondiskVersion = 1;</code> 855 * 856 * <pre> 857 * The version of the above EBNF grammars. 858 * </pre> 859 */ 860 public boolean hasOndiskVersion() { 861 return ((bitField0_ & 0x00000001) == 0x00000001); 862 } 863 /** 864 * <code>required uint32 ondiskVersion = 1;</code> 865 * 866 * <pre> 867 * The version of the above EBNF grammars. 868 * </pre> 869 */ 870 public int getOndiskVersion() { 871 return ondiskVersion_; 872 } 873 874 // required uint32 layoutVersion = 2; 875 public static final int LAYOUTVERSION_FIELD_NUMBER = 2; 876 private int layoutVersion_; 877 /** 878 * <code>required uint32 layoutVersion = 2;</code> 879 * 880 * <pre> 881 * layoutVersion describes which features are available in the 882 * FSImage. 883 * </pre> 884 */ 885 public boolean hasLayoutVersion() { 886 return ((bitField0_ & 0x00000002) == 0x00000002); 887 } 888 /** 889 * <code>required uint32 layoutVersion = 2;</code> 890 * 891 * <pre> 892 * layoutVersion describes which features are available in the 893 * FSImage. 894 * </pre> 895 */ 896 public int getLayoutVersion() { 897 return layoutVersion_; 898 } 899 900 // optional string codec = 3; 901 public static final int CODEC_FIELD_NUMBER = 3; 902 private java.lang.Object codec_; 903 /** 904 * <code>optional string codec = 3;</code> 905 */ 906 public boolean hasCodec() { 907 return ((bitField0_ & 0x00000004) == 0x00000004); 908 } 909 /** 910 * <code>optional string codec = 3;</code> 911 */ 912 public java.lang.String getCodec() { 913 java.lang.Object ref = codec_; 914 if (ref instanceof java.lang.String) { 915 return (java.lang.String) ref; 916 } else { 917 com.google.protobuf.ByteString bs = 918 (com.google.protobuf.ByteString) ref; 919 java.lang.String s = bs.toStringUtf8(); 920 if (bs.isValidUtf8()) { 921 codec_ = s; 922 } 923 return s; 924 } 925 } 926 /** 927 * <code>optional string codec = 3;</code> 928 */ 929 public com.google.protobuf.ByteString 930 getCodecBytes() { 931 java.lang.Object ref = codec_; 932 if (ref instanceof java.lang.String) { 933 com.google.protobuf.ByteString b = 934 com.google.protobuf.ByteString.copyFromUtf8( 935 (java.lang.String) ref); 936 codec_ = b; 937 return b; 938 } else { 939 return (com.google.protobuf.ByteString) ref; 940 } 941 } 942 943 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 944 public static final int SECTIONS_FIELD_NUMBER = 4; 945 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> sections_; 946 /** 947 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 948 */ 949 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> getSectionsList() { 950 return sections_; 951 } 952 /** 953 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 954 */ 955 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 956 getSectionsOrBuilderList() { 957 return sections_; 958 } 959 /** 960 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 961 */ 962 public int getSectionsCount() { 963 return sections_.size(); 964 } 965 /** 966 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 967 */ 968 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index) { 969 return sections_.get(index); 970 } 971 /** 972 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 973 */ 974 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 975 int index) { 976 return sections_.get(index); 977 } 978 979 private void initFields() { 980 ondiskVersion_ = 0; 981 layoutVersion_ = 0; 982 codec_ = ""; 983 sections_ = java.util.Collections.emptyList(); 984 } 985 private byte memoizedIsInitialized = -1; 986 public final boolean isInitialized() { 987 byte isInitialized = memoizedIsInitialized; 988 if (isInitialized != -1) return isInitialized == 1; 989 990 if (!hasOndiskVersion()) { 991 memoizedIsInitialized = 0; 992 return false; 993 } 994 if (!hasLayoutVersion()) { 995 memoizedIsInitialized = 0; 996 return false; 997 } 998 memoizedIsInitialized = 1; 999 return true; 1000 } 1001 1002 public void writeTo(com.google.protobuf.CodedOutputStream output) 1003 throws java.io.IOException { 1004 getSerializedSize(); 1005 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1006 output.writeUInt32(1, ondiskVersion_); 1007 } 1008 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1009 output.writeUInt32(2, layoutVersion_); 1010 } 1011 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1012 output.writeBytes(3, getCodecBytes()); 1013 } 1014 for (int i = 0; i < sections_.size(); i++) { 1015 output.writeMessage(4, sections_.get(i)); 1016 } 1017 getUnknownFields().writeTo(output); 1018 } 1019 1020 private int memoizedSerializedSize = -1; 1021 public int getSerializedSize() { 1022 int size = memoizedSerializedSize; 1023 if (size != -1) return size; 1024 1025 size = 0; 1026 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1027 size += com.google.protobuf.CodedOutputStream 1028 .computeUInt32Size(1, ondiskVersion_); 1029 } 1030 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1031 size += com.google.protobuf.CodedOutputStream 1032 .computeUInt32Size(2, layoutVersion_); 1033 } 1034 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1035 size += com.google.protobuf.CodedOutputStream 1036 .computeBytesSize(3, getCodecBytes()); 1037 } 1038 for (int i = 0; i < sections_.size(); i++) { 1039 size += com.google.protobuf.CodedOutputStream 1040 .computeMessageSize(4, sections_.get(i)); 1041 } 1042 size += getUnknownFields().getSerializedSize(); 1043 memoizedSerializedSize = size; 1044 return size; 1045 } 1046 1047 private static final long serialVersionUID = 0L; 1048 @java.lang.Override 1049 protected java.lang.Object writeReplace() 1050 throws java.io.ObjectStreamException { 1051 return super.writeReplace(); 1052 } 1053 1054 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1055 com.google.protobuf.ByteString data) 1056 throws com.google.protobuf.InvalidProtocolBufferException { 1057 return PARSER.parseFrom(data); 1058 } 1059 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1060 com.google.protobuf.ByteString data, 1061 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1062 throws com.google.protobuf.InvalidProtocolBufferException { 1063 return PARSER.parseFrom(data, extensionRegistry); 1064 } 1065 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom(byte[] data) 1066 throws com.google.protobuf.InvalidProtocolBufferException { 1067 return PARSER.parseFrom(data); 1068 } 1069 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1070 byte[] data, 1071 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1072 throws com.google.protobuf.InvalidProtocolBufferException { 1073 return PARSER.parseFrom(data, extensionRegistry); 1074 } 1075 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom(java.io.InputStream input) 1076 throws java.io.IOException { 1077 return PARSER.parseFrom(input); 1078 } 1079 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1080 java.io.InputStream input, 1081 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1082 throws java.io.IOException { 1083 return PARSER.parseFrom(input, extensionRegistry); 1084 } 1085 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseDelimitedFrom(java.io.InputStream input) 1086 throws java.io.IOException { 1087 return PARSER.parseDelimitedFrom(input); 1088 } 1089 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseDelimitedFrom( 1090 java.io.InputStream input, 1091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1092 throws java.io.IOException { 1093 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1094 } 1095 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1096 com.google.protobuf.CodedInputStream input) 1097 throws java.io.IOException { 1098 return PARSER.parseFrom(input); 1099 } 1100 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1101 com.google.protobuf.CodedInputStream input, 1102 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1103 throws java.io.IOException { 1104 return PARSER.parseFrom(input, extensionRegistry); 1105 } 1106 1107 public static Builder newBuilder() { return Builder.create(); } 1108 public Builder newBuilderForType() { return newBuilder(); } 1109 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary prototype) { 1110 return newBuilder().mergeFrom(prototype); 1111 } 1112 public Builder toBuilder() { return newBuilder(this); } 1113 1114 @java.lang.Override 1115 protected Builder newBuilderForType( 1116 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1117 Builder builder = new Builder(parent); 1118 return builder; 1119 } 1120 /** 1121 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary} 1122 */ 1123 public static final class Builder extends 1124 com.google.protobuf.GeneratedMessage.Builder<Builder> 1125 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummaryOrBuilder { 1126 public static final com.google.protobuf.Descriptors.Descriptor 1127 getDescriptor() { 1128 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 1129 } 1130 1131 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1132 internalGetFieldAccessorTable() { 1133 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable 1134 .ensureFieldAccessorsInitialized( 1135 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Builder.class); 1136 } 1137 1138 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.newBuilder() 1139 private Builder() { 1140 maybeForceBuilderInitialization(); 1141 } 1142 1143 private Builder( 1144 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1145 super(parent); 1146 maybeForceBuilderInitialization(); 1147 } 1148 private void maybeForceBuilderInitialization() { 1149 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1150 getSectionsFieldBuilder(); 1151 } 1152 } 1153 private static Builder create() { 1154 return new Builder(); 1155 } 1156 1157 public Builder clear() { 1158 super.clear(); 1159 ondiskVersion_ = 0; 1160 bitField0_ = (bitField0_ & ~0x00000001); 1161 layoutVersion_ = 0; 1162 bitField0_ = (bitField0_ & ~0x00000002); 1163 codec_ = ""; 1164 bitField0_ = (bitField0_ & ~0x00000004); 1165 if (sectionsBuilder_ == null) { 1166 sections_ = java.util.Collections.emptyList(); 1167 bitField0_ = (bitField0_ & ~0x00000008); 1168 } else { 1169 sectionsBuilder_.clear(); 1170 } 1171 return this; 1172 } 1173 1174 public Builder clone() { 1175 return create().mergeFrom(buildPartial()); 1176 } 1177 1178 public com.google.protobuf.Descriptors.Descriptor 1179 getDescriptorForType() { 1180 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 1181 } 1182 1183 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary getDefaultInstanceForType() { 1184 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.getDefaultInstance(); 1185 } 1186 1187 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary build() { 1188 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary result = buildPartial(); 1189 if (!result.isInitialized()) { 1190 throw newUninitializedMessageException(result); 1191 } 1192 return result; 1193 } 1194 1195 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary buildPartial() { 1196 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary(this); 1197 int from_bitField0_ = bitField0_; 1198 int to_bitField0_ = 0; 1199 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1200 to_bitField0_ |= 0x00000001; 1201 } 1202 result.ondiskVersion_ = ondiskVersion_; 1203 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 1204 to_bitField0_ |= 0x00000002; 1205 } 1206 result.layoutVersion_ = layoutVersion_; 1207 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 1208 to_bitField0_ |= 0x00000004; 1209 } 1210 result.codec_ = codec_; 1211 if (sectionsBuilder_ == null) { 1212 if (((bitField0_ & 0x00000008) == 0x00000008)) { 1213 sections_ = java.util.Collections.unmodifiableList(sections_); 1214 bitField0_ = (bitField0_ & ~0x00000008); 1215 } 1216 result.sections_ = sections_; 1217 } else { 1218 result.sections_ = sectionsBuilder_.build(); 1219 } 1220 result.bitField0_ = to_bitField0_; 1221 onBuilt(); 1222 return result; 1223 } 1224 1225 public Builder mergeFrom(com.google.protobuf.Message other) { 1226 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary) { 1227 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary)other); 1228 } else { 1229 super.mergeFrom(other); 1230 return this; 1231 } 1232 } 1233 1234 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary other) { 1235 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.getDefaultInstance()) return this; 1236 if (other.hasOndiskVersion()) { 1237 setOndiskVersion(other.getOndiskVersion()); 1238 } 1239 if (other.hasLayoutVersion()) { 1240 setLayoutVersion(other.getLayoutVersion()); 1241 } 1242 if (other.hasCodec()) { 1243 bitField0_ |= 0x00000004; 1244 codec_ = other.codec_; 1245 onChanged(); 1246 } 1247 if (sectionsBuilder_ == null) { 1248 if (!other.sections_.isEmpty()) { 1249 if (sections_.isEmpty()) { 1250 sections_ = other.sections_; 1251 bitField0_ = (bitField0_ & ~0x00000008); 1252 } else { 1253 ensureSectionsIsMutable(); 1254 sections_.addAll(other.sections_); 1255 } 1256 onChanged(); 1257 } 1258 } else { 1259 if (!other.sections_.isEmpty()) { 1260 if (sectionsBuilder_.isEmpty()) { 1261 sectionsBuilder_.dispose(); 1262 sectionsBuilder_ = null; 1263 sections_ = other.sections_; 1264 bitField0_ = (bitField0_ & ~0x00000008); 1265 sectionsBuilder_ = 1266 com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? 1267 getSectionsFieldBuilder() : null; 1268 } else { 1269 sectionsBuilder_.addAllMessages(other.sections_); 1270 } 1271 } 1272 } 1273 this.mergeUnknownFields(other.getUnknownFields()); 1274 return this; 1275 } 1276 1277 public final boolean isInitialized() { 1278 if (!hasOndiskVersion()) { 1279 1280 return false; 1281 } 1282 if (!hasLayoutVersion()) { 1283 1284 return false; 1285 } 1286 return true; 1287 } 1288 1289 public Builder mergeFrom( 1290 com.google.protobuf.CodedInputStream input, 1291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1292 throws java.io.IOException { 1293 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parsedMessage = null; 1294 try { 1295 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1296 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1297 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary) e.getUnfinishedMessage(); 1298 throw e; 1299 } finally { 1300 if (parsedMessage != null) { 1301 mergeFrom(parsedMessage); 1302 } 1303 } 1304 return this; 1305 } 1306 private int bitField0_; 1307 1308 // required uint32 ondiskVersion = 1; 1309 private int ondiskVersion_ ; 1310 /** 1311 * <code>required uint32 ondiskVersion = 1;</code> 1312 * 1313 * <pre> 1314 * The version of the above EBNF grammars. 1315 * </pre> 1316 */ 1317 public boolean hasOndiskVersion() { 1318 return ((bitField0_ & 0x00000001) == 0x00000001); 1319 } 1320 /** 1321 * <code>required uint32 ondiskVersion = 1;</code> 1322 * 1323 * <pre> 1324 * The version of the above EBNF grammars. 1325 * </pre> 1326 */ 1327 public int getOndiskVersion() { 1328 return ondiskVersion_; 1329 } 1330 /** 1331 * <code>required uint32 ondiskVersion = 1;</code> 1332 * 1333 * <pre> 1334 * The version of the above EBNF grammars. 1335 * </pre> 1336 */ 1337 public Builder setOndiskVersion(int value) { 1338 bitField0_ |= 0x00000001; 1339 ondiskVersion_ = value; 1340 onChanged(); 1341 return this; 1342 } 1343 /** 1344 * <code>required uint32 ondiskVersion = 1;</code> 1345 * 1346 * <pre> 1347 * The version of the above EBNF grammars. 1348 * </pre> 1349 */ 1350 public Builder clearOndiskVersion() { 1351 bitField0_ = (bitField0_ & ~0x00000001); 1352 ondiskVersion_ = 0; 1353 onChanged(); 1354 return this; 1355 } 1356 1357 // required uint32 layoutVersion = 2; 1358 private int layoutVersion_ ; 1359 /** 1360 * <code>required uint32 layoutVersion = 2;</code> 1361 * 1362 * <pre> 1363 * layoutVersion describes which features are available in the 1364 * FSImage. 1365 * </pre> 1366 */ 1367 public boolean hasLayoutVersion() { 1368 return ((bitField0_ & 0x00000002) == 0x00000002); 1369 } 1370 /** 1371 * <code>required uint32 layoutVersion = 2;</code> 1372 * 1373 * <pre> 1374 * layoutVersion describes which features are available in the 1375 * FSImage. 1376 * </pre> 1377 */ 1378 public int getLayoutVersion() { 1379 return layoutVersion_; 1380 } 1381 /** 1382 * <code>required uint32 layoutVersion = 2;</code> 1383 * 1384 * <pre> 1385 * layoutVersion describes which features are available in the 1386 * FSImage. 1387 * </pre> 1388 */ 1389 public Builder setLayoutVersion(int value) { 1390 bitField0_ |= 0x00000002; 1391 layoutVersion_ = value; 1392 onChanged(); 1393 return this; 1394 } 1395 /** 1396 * <code>required uint32 layoutVersion = 2;</code> 1397 * 1398 * <pre> 1399 * layoutVersion describes which features are available in the 1400 * FSImage. 1401 * </pre> 1402 */ 1403 public Builder clearLayoutVersion() { 1404 bitField0_ = (bitField0_ & ~0x00000002); 1405 layoutVersion_ = 0; 1406 onChanged(); 1407 return this; 1408 } 1409 1410 // optional string codec = 3; 1411 private java.lang.Object codec_ = ""; 1412 /** 1413 * <code>optional string codec = 3;</code> 1414 */ 1415 public boolean hasCodec() { 1416 return ((bitField0_ & 0x00000004) == 0x00000004); 1417 } 1418 /** 1419 * <code>optional string codec = 3;</code> 1420 */ 1421 public java.lang.String getCodec() { 1422 java.lang.Object ref = codec_; 1423 if (!(ref instanceof java.lang.String)) { 1424 java.lang.String s = ((com.google.protobuf.ByteString) ref) 1425 .toStringUtf8(); 1426 codec_ = s; 1427 return s; 1428 } else { 1429 return (java.lang.String) ref; 1430 } 1431 } 1432 /** 1433 * <code>optional string codec = 3;</code> 1434 */ 1435 public com.google.protobuf.ByteString 1436 getCodecBytes() { 1437 java.lang.Object ref = codec_; 1438 if (ref instanceof String) { 1439 com.google.protobuf.ByteString b = 1440 com.google.protobuf.ByteString.copyFromUtf8( 1441 (java.lang.String) ref); 1442 codec_ = b; 1443 return b; 1444 } else { 1445 return (com.google.protobuf.ByteString) ref; 1446 } 1447 } 1448 /** 1449 * <code>optional string codec = 3;</code> 1450 */ 1451 public Builder setCodec( 1452 java.lang.String value) { 1453 if (value == null) { 1454 throw new NullPointerException(); 1455 } 1456 bitField0_ |= 0x00000004; 1457 codec_ = value; 1458 onChanged(); 1459 return this; 1460 } 1461 /** 1462 * <code>optional string codec = 3;</code> 1463 */ 1464 public Builder clearCodec() { 1465 bitField0_ = (bitField0_ & ~0x00000004); 1466 codec_ = getDefaultInstance().getCodec(); 1467 onChanged(); 1468 return this; 1469 } 1470 /** 1471 * <code>optional string codec = 3;</code> 1472 */ 1473 public Builder setCodecBytes( 1474 com.google.protobuf.ByteString value) { 1475 if (value == null) { 1476 throw new NullPointerException(); 1477 } 1478 bitField0_ |= 0x00000004; 1479 codec_ = value; 1480 onChanged(); 1481 return this; 1482 } 1483 1484 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 1485 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> sections_ = 1486 java.util.Collections.emptyList(); 1487 private void ensureSectionsIsMutable() { 1488 if (!((bitField0_ & 0x00000008) == 0x00000008)) { 1489 sections_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section>(sections_); 1490 bitField0_ |= 0x00000008; 1491 } 1492 } 1493 1494 private com.google.protobuf.RepeatedFieldBuilder< 1495 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> sectionsBuilder_; 1496 1497 /** 1498 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1499 */ 1500 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> getSectionsList() { 1501 if (sectionsBuilder_ == null) { 1502 return java.util.Collections.unmodifiableList(sections_); 1503 } else { 1504 return sectionsBuilder_.getMessageList(); 1505 } 1506 } 1507 /** 1508 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1509 */ 1510 public int getSectionsCount() { 1511 if (sectionsBuilder_ == null) { 1512 return sections_.size(); 1513 } else { 1514 return sectionsBuilder_.getCount(); 1515 } 1516 } 1517 /** 1518 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1519 */ 1520 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index) { 1521 if (sectionsBuilder_ == null) { 1522 return sections_.get(index); 1523 } else { 1524 return sectionsBuilder_.getMessage(index); 1525 } 1526 } 1527 /** 1528 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1529 */ 1530 public Builder setSections( 1531 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1532 if (sectionsBuilder_ == null) { 1533 if (value == null) { 1534 throw new NullPointerException(); 1535 } 1536 ensureSectionsIsMutable(); 1537 sections_.set(index, value); 1538 onChanged(); 1539 } else { 1540 sectionsBuilder_.setMessage(index, value); 1541 } 1542 return this; 1543 } 1544 /** 1545 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1546 */ 1547 public Builder setSections( 1548 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1549 if (sectionsBuilder_ == null) { 1550 ensureSectionsIsMutable(); 1551 sections_.set(index, builderForValue.build()); 1552 onChanged(); 1553 } else { 1554 sectionsBuilder_.setMessage(index, builderForValue.build()); 1555 } 1556 return this; 1557 } 1558 /** 1559 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1560 */ 1561 public Builder addSections(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1562 if (sectionsBuilder_ == null) { 1563 if (value == null) { 1564 throw new NullPointerException(); 1565 } 1566 ensureSectionsIsMutable(); 1567 sections_.add(value); 1568 onChanged(); 1569 } else { 1570 sectionsBuilder_.addMessage(value); 1571 } 1572 return this; 1573 } 1574 /** 1575 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1576 */ 1577 public Builder addSections( 1578 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1579 if (sectionsBuilder_ == null) { 1580 if (value == null) { 1581 throw new NullPointerException(); 1582 } 1583 ensureSectionsIsMutable(); 1584 sections_.add(index, value); 1585 onChanged(); 1586 } else { 1587 sectionsBuilder_.addMessage(index, value); 1588 } 1589 return this; 1590 } 1591 /** 1592 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1593 */ 1594 public Builder addSections( 1595 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1596 if (sectionsBuilder_ == null) { 1597 ensureSectionsIsMutable(); 1598 sections_.add(builderForValue.build()); 1599 onChanged(); 1600 } else { 1601 sectionsBuilder_.addMessage(builderForValue.build()); 1602 } 1603 return this; 1604 } 1605 /** 1606 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1607 */ 1608 public Builder addSections( 1609 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1610 if (sectionsBuilder_ == null) { 1611 ensureSectionsIsMutable(); 1612 sections_.add(index, builderForValue.build()); 1613 onChanged(); 1614 } else { 1615 sectionsBuilder_.addMessage(index, builderForValue.build()); 1616 } 1617 return this; 1618 } 1619 /** 1620 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1621 */ 1622 public Builder addAllSections( 1623 java.lang.Iterable<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> values) { 1624 if (sectionsBuilder_ == null) { 1625 ensureSectionsIsMutable(); 1626 super.addAll(values, sections_); 1627 onChanged(); 1628 } else { 1629 sectionsBuilder_.addAllMessages(values); 1630 } 1631 return this; 1632 } 1633 /** 1634 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1635 */ 1636 public Builder clearSections() { 1637 if (sectionsBuilder_ == null) { 1638 sections_ = java.util.Collections.emptyList(); 1639 bitField0_ = (bitField0_ & ~0x00000008); 1640 onChanged(); 1641 } else { 1642 sectionsBuilder_.clear(); 1643 } 1644 return this; 1645 } 1646 /** 1647 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1648 */ 1649 public Builder removeSections(int index) { 1650 if (sectionsBuilder_ == null) { 1651 ensureSectionsIsMutable(); 1652 sections_.remove(index); 1653 onChanged(); 1654 } else { 1655 sectionsBuilder_.remove(index); 1656 } 1657 return this; 1658 } 1659 /** 1660 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1661 */ 1662 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder getSectionsBuilder( 1663 int index) { 1664 return getSectionsFieldBuilder().getBuilder(index); 1665 } 1666 /** 1667 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1668 */ 1669 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 1670 int index) { 1671 if (sectionsBuilder_ == null) { 1672 return sections_.get(index); } else { 1673 return sectionsBuilder_.getMessageOrBuilder(index); 1674 } 1675 } 1676 /** 1677 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1678 */ 1679 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 1680 getSectionsOrBuilderList() { 1681 if (sectionsBuilder_ != null) { 1682 return sectionsBuilder_.getMessageOrBuilderList(); 1683 } else { 1684 return java.util.Collections.unmodifiableList(sections_); 1685 } 1686 } 1687 /** 1688 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1689 */ 1690 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder addSectionsBuilder() { 1691 return getSectionsFieldBuilder().addBuilder( 1692 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()); 1693 } 1694 /** 1695 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1696 */ 1697 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder addSectionsBuilder( 1698 int index) { 1699 return getSectionsFieldBuilder().addBuilder( 1700 index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()); 1701 } 1702 /** 1703 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1704 */ 1705 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder> 1706 getSectionsBuilderList() { 1707 return getSectionsFieldBuilder().getBuilderList(); 1708 } 1709 private com.google.protobuf.RepeatedFieldBuilder< 1710 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 1711 getSectionsFieldBuilder() { 1712 if (sectionsBuilder_ == null) { 1713 sectionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< 1714 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder>( 1715 sections_, 1716 ((bitField0_ & 0x00000008) == 0x00000008), 1717 getParentForChildren(), 1718 isClean()); 1719 sections_ = null; 1720 } 1721 return sectionsBuilder_; 1722 } 1723 1724 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FileSummary) 1725 } 1726 1727 static { 1728 defaultInstance = new FileSummary(true); 1729 defaultInstance.initFields(); 1730 } 1731 1732 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FileSummary) 1733 } 1734 1735 public interface NameSystemSectionOrBuilder 1736 extends com.google.protobuf.MessageOrBuilder { 1737 1738 // optional uint32 namespaceId = 1; 1739 /** 1740 * <code>optional uint32 namespaceId = 1;</code> 1741 */ 1742 boolean hasNamespaceId(); 1743 /** 1744 * <code>optional uint32 namespaceId = 1;</code> 1745 */ 1746 int getNamespaceId(); 1747 1748 // optional uint64 genstampV1 = 2; 1749 /** 1750 * <code>optional uint64 genstampV1 = 2;</code> 1751 */ 1752 boolean hasGenstampV1(); 1753 /** 1754 * <code>optional uint64 genstampV1 = 2;</code> 1755 */ 1756 long getGenstampV1(); 1757 1758 // optional uint64 genstampV2 = 3; 1759 /** 1760 * <code>optional uint64 genstampV2 = 3;</code> 1761 */ 1762 boolean hasGenstampV2(); 1763 /** 1764 * <code>optional uint64 genstampV2 = 3;</code> 1765 */ 1766 long getGenstampV2(); 1767 1768 // optional uint64 genstampV1Limit = 4; 1769 /** 1770 * <code>optional uint64 genstampV1Limit = 4;</code> 1771 */ 1772 boolean hasGenstampV1Limit(); 1773 /** 1774 * <code>optional uint64 genstampV1Limit = 4;</code> 1775 */ 1776 long getGenstampV1Limit(); 1777 1778 // optional uint64 lastAllocatedBlockId = 5; 1779 /** 1780 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 1781 */ 1782 boolean hasLastAllocatedBlockId(); 1783 /** 1784 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 1785 */ 1786 long getLastAllocatedBlockId(); 1787 1788 // optional uint64 transactionId = 6; 1789 /** 1790 * <code>optional uint64 transactionId = 6;</code> 1791 */ 1792 boolean hasTransactionId(); 1793 /** 1794 * <code>optional uint64 transactionId = 6;</code> 1795 */ 1796 long getTransactionId(); 1797 1798 // optional uint64 rollingUpgradeStartTime = 7; 1799 /** 1800 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 1801 */ 1802 boolean hasRollingUpgradeStartTime(); 1803 /** 1804 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 1805 */ 1806 long getRollingUpgradeStartTime(); 1807 } 1808 /** 1809 * Protobuf type {@code hadoop.hdfs.fsimage.NameSystemSection} 1810 * 1811 * <pre> 1812 ** 1813 * Name: NS_INFO 1814 * </pre> 1815 */ 1816 public static final class NameSystemSection extends 1817 com.google.protobuf.GeneratedMessage 1818 implements NameSystemSectionOrBuilder { 1819 // Use NameSystemSection.newBuilder() to construct. 1820 private NameSystemSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1821 super(builder); 1822 this.unknownFields = builder.getUnknownFields(); 1823 } 1824 private NameSystemSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1825 1826 private static final NameSystemSection defaultInstance; 1827 public static NameSystemSection getDefaultInstance() { 1828 return defaultInstance; 1829 } 1830 1831 public NameSystemSection getDefaultInstanceForType() { 1832 return defaultInstance; 1833 } 1834 1835 private final com.google.protobuf.UnknownFieldSet unknownFields; 1836 @java.lang.Override 1837 public final com.google.protobuf.UnknownFieldSet 1838 getUnknownFields() { 1839 return this.unknownFields; 1840 } 1841 private NameSystemSection( 1842 com.google.protobuf.CodedInputStream input, 1843 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1844 throws com.google.protobuf.InvalidProtocolBufferException { 1845 initFields(); 1846 int mutable_bitField0_ = 0; 1847 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1848 com.google.protobuf.UnknownFieldSet.newBuilder(); 1849 try { 1850 boolean done = false; 1851 while (!done) { 1852 int tag = input.readTag(); 1853 switch (tag) { 1854 case 0: 1855 done = true; 1856 break; 1857 default: { 1858 if (!parseUnknownField(input, unknownFields, 1859 extensionRegistry, tag)) { 1860 done = true; 1861 } 1862 break; 1863 } 1864 case 8: { 1865 bitField0_ |= 0x00000001; 1866 namespaceId_ = input.readUInt32(); 1867 break; 1868 } 1869 case 16: { 1870 bitField0_ |= 0x00000002; 1871 genstampV1_ = input.readUInt64(); 1872 break; 1873 } 1874 case 24: { 1875 bitField0_ |= 0x00000004; 1876 genstampV2_ = input.readUInt64(); 1877 break; 1878 } 1879 case 32: { 1880 bitField0_ |= 0x00000008; 1881 genstampV1Limit_ = input.readUInt64(); 1882 break; 1883 } 1884 case 40: { 1885 bitField0_ |= 0x00000010; 1886 lastAllocatedBlockId_ = input.readUInt64(); 1887 break; 1888 } 1889 case 48: { 1890 bitField0_ |= 0x00000020; 1891 transactionId_ = input.readUInt64(); 1892 break; 1893 } 1894 case 56: { 1895 bitField0_ |= 0x00000040; 1896 rollingUpgradeStartTime_ = input.readUInt64(); 1897 break; 1898 } 1899 } 1900 } 1901 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1902 throw e.setUnfinishedMessage(this); 1903 } catch (java.io.IOException e) { 1904 throw new com.google.protobuf.InvalidProtocolBufferException( 1905 e.getMessage()).setUnfinishedMessage(this); 1906 } finally { 1907 this.unknownFields = unknownFields.build(); 1908 makeExtensionsImmutable(); 1909 } 1910 } 1911 public static final com.google.protobuf.Descriptors.Descriptor 1912 getDescriptor() { 1913 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 1914 } 1915 1916 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1917 internalGetFieldAccessorTable() { 1918 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable 1919 .ensureFieldAccessorsInitialized( 1920 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.Builder.class); 1921 } 1922 1923 public static com.google.protobuf.Parser<NameSystemSection> PARSER = 1924 new com.google.protobuf.AbstractParser<NameSystemSection>() { 1925 public NameSystemSection parsePartialFrom( 1926 com.google.protobuf.CodedInputStream input, 1927 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1928 throws com.google.protobuf.InvalidProtocolBufferException { 1929 return new NameSystemSection(input, extensionRegistry); 1930 } 1931 }; 1932 1933 @java.lang.Override 1934 public com.google.protobuf.Parser<NameSystemSection> getParserForType() { 1935 return PARSER; 1936 } 1937 1938 private int bitField0_; 1939 // optional uint32 namespaceId = 1; 1940 public static final int NAMESPACEID_FIELD_NUMBER = 1; 1941 private int namespaceId_; 1942 /** 1943 * <code>optional uint32 namespaceId = 1;</code> 1944 */ 1945 public boolean hasNamespaceId() { 1946 return ((bitField0_ & 0x00000001) == 0x00000001); 1947 } 1948 /** 1949 * <code>optional uint32 namespaceId = 1;</code> 1950 */ 1951 public int getNamespaceId() { 1952 return namespaceId_; 1953 } 1954 1955 // optional uint64 genstampV1 = 2; 1956 public static final int GENSTAMPV1_FIELD_NUMBER = 2; 1957 private long genstampV1_; 1958 /** 1959 * <code>optional uint64 genstampV1 = 2;</code> 1960 */ 1961 public boolean hasGenstampV1() { 1962 return ((bitField0_ & 0x00000002) == 0x00000002); 1963 } 1964 /** 1965 * <code>optional uint64 genstampV1 = 2;</code> 1966 */ 1967 public long getGenstampV1() { 1968 return genstampV1_; 1969 } 1970 1971 // optional uint64 genstampV2 = 3; 1972 public static final int GENSTAMPV2_FIELD_NUMBER = 3; 1973 private long genstampV2_; 1974 /** 1975 * <code>optional uint64 genstampV2 = 3;</code> 1976 */ 1977 public boolean hasGenstampV2() { 1978 return ((bitField0_ & 0x00000004) == 0x00000004); 1979 } 1980 /** 1981 * <code>optional uint64 genstampV2 = 3;</code> 1982 */ 1983 public long getGenstampV2() { 1984 return genstampV2_; 1985 } 1986 1987 // optional uint64 genstampV1Limit = 4; 1988 public static final int GENSTAMPV1LIMIT_FIELD_NUMBER = 4; 1989 private long genstampV1Limit_; 1990 /** 1991 * <code>optional uint64 genstampV1Limit = 4;</code> 1992 */ 1993 public boolean hasGenstampV1Limit() { 1994 return ((bitField0_ & 0x00000008) == 0x00000008); 1995 } 1996 /** 1997 * <code>optional uint64 genstampV1Limit = 4;</code> 1998 */ 1999 public long getGenstampV1Limit() { 2000 return genstampV1Limit_; 2001 } 2002 2003 // optional uint64 lastAllocatedBlockId = 5; 2004 public static final int LASTALLOCATEDBLOCKID_FIELD_NUMBER = 5; 2005 private long lastAllocatedBlockId_; 2006 /** 2007 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2008 */ 2009 public boolean hasLastAllocatedBlockId() { 2010 return ((bitField0_ & 0x00000010) == 0x00000010); 2011 } 2012 /** 2013 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2014 */ 2015 public long getLastAllocatedBlockId() { 2016 return lastAllocatedBlockId_; 2017 } 2018 2019 // optional uint64 transactionId = 6; 2020 public static final int TRANSACTIONID_FIELD_NUMBER = 6; 2021 private long transactionId_; 2022 /** 2023 * <code>optional uint64 transactionId = 6;</code> 2024 */ 2025 public boolean hasTransactionId() { 2026 return ((bitField0_ & 0x00000020) == 0x00000020); 2027 } 2028 /** 2029 * <code>optional uint64 transactionId = 6;</code> 2030 */ 2031 public long getTransactionId() { 2032 return transactionId_; 2033 } 2034 2035 // optional uint64 rollingUpgradeStartTime = 7; 2036 public static final int ROLLINGUPGRADESTARTTIME_FIELD_NUMBER = 7; 2037 private long rollingUpgradeStartTime_; 2038 /** 2039 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2040 */ 2041 public boolean hasRollingUpgradeStartTime() { 2042 return ((bitField0_ & 0x00000040) == 0x00000040); 2043 } 2044 /** 2045 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2046 */ 2047 public long getRollingUpgradeStartTime() { 2048 return rollingUpgradeStartTime_; 2049 } 2050 2051 private void initFields() { 2052 namespaceId_ = 0; 2053 genstampV1_ = 0L; 2054 genstampV2_ = 0L; 2055 genstampV1Limit_ = 0L; 2056 lastAllocatedBlockId_ = 0L; 2057 transactionId_ = 0L; 2058 rollingUpgradeStartTime_ = 0L; 2059 } 2060 private byte memoizedIsInitialized = -1; 2061 public final boolean isInitialized() { 2062 byte isInitialized = memoizedIsInitialized; 2063 if (isInitialized != -1) return isInitialized == 1; 2064 2065 memoizedIsInitialized = 1; 2066 return true; 2067 } 2068 2069 public void writeTo(com.google.protobuf.CodedOutputStream output) 2070 throws java.io.IOException { 2071 getSerializedSize(); 2072 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2073 output.writeUInt32(1, namespaceId_); 2074 } 2075 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2076 output.writeUInt64(2, genstampV1_); 2077 } 2078 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2079 output.writeUInt64(3, genstampV2_); 2080 } 2081 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2082 output.writeUInt64(4, genstampV1Limit_); 2083 } 2084 if (((bitField0_ & 0x00000010) == 0x00000010)) { 2085 output.writeUInt64(5, lastAllocatedBlockId_); 2086 } 2087 if (((bitField0_ & 0x00000020) == 0x00000020)) { 2088 output.writeUInt64(6, transactionId_); 2089 } 2090 if (((bitField0_ & 0x00000040) == 0x00000040)) { 2091 output.writeUInt64(7, rollingUpgradeStartTime_); 2092 } 2093 getUnknownFields().writeTo(output); 2094 } 2095 2096 private int memoizedSerializedSize = -1; 2097 public int getSerializedSize() { 2098 int size = memoizedSerializedSize; 2099 if (size != -1) return size; 2100 2101 size = 0; 2102 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2103 size += com.google.protobuf.CodedOutputStream 2104 .computeUInt32Size(1, namespaceId_); 2105 } 2106 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2107 size += com.google.protobuf.CodedOutputStream 2108 .computeUInt64Size(2, genstampV1_); 2109 } 2110 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2111 size += com.google.protobuf.CodedOutputStream 2112 .computeUInt64Size(3, genstampV2_); 2113 } 2114 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2115 size += com.google.protobuf.CodedOutputStream 2116 .computeUInt64Size(4, genstampV1Limit_); 2117 } 2118 if (((bitField0_ & 0x00000010) == 0x00000010)) { 2119 size += com.google.protobuf.CodedOutputStream 2120 .computeUInt64Size(5, lastAllocatedBlockId_); 2121 } 2122 if (((bitField0_ & 0x00000020) == 0x00000020)) { 2123 size += com.google.protobuf.CodedOutputStream 2124 .computeUInt64Size(6, transactionId_); 2125 } 2126 if (((bitField0_ & 0x00000040) == 0x00000040)) { 2127 size += com.google.protobuf.CodedOutputStream 2128 .computeUInt64Size(7, rollingUpgradeStartTime_); 2129 } 2130 size += getUnknownFields().getSerializedSize(); 2131 memoizedSerializedSize = size; 2132 return size; 2133 } 2134 2135 private static final long serialVersionUID = 0L; 2136 @java.lang.Override 2137 protected java.lang.Object writeReplace() 2138 throws java.io.ObjectStreamException { 2139 return super.writeReplace(); 2140 } 2141 2142 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2143 com.google.protobuf.ByteString data) 2144 throws com.google.protobuf.InvalidProtocolBufferException { 2145 return PARSER.parseFrom(data); 2146 } 2147 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2148 com.google.protobuf.ByteString data, 2149 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2150 throws com.google.protobuf.InvalidProtocolBufferException { 2151 return PARSER.parseFrom(data, extensionRegistry); 2152 } 2153 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom(byte[] data) 2154 throws com.google.protobuf.InvalidProtocolBufferException { 2155 return PARSER.parseFrom(data); 2156 } 2157 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2158 byte[] data, 2159 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2160 throws com.google.protobuf.InvalidProtocolBufferException { 2161 return PARSER.parseFrom(data, extensionRegistry); 2162 } 2163 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom(java.io.InputStream input) 2164 throws java.io.IOException { 2165 return PARSER.parseFrom(input); 2166 } 2167 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2168 java.io.InputStream input, 2169 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2170 throws java.io.IOException { 2171 return PARSER.parseFrom(input, extensionRegistry); 2172 } 2173 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseDelimitedFrom(java.io.InputStream input) 2174 throws java.io.IOException { 2175 return PARSER.parseDelimitedFrom(input); 2176 } 2177 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseDelimitedFrom( 2178 java.io.InputStream input, 2179 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2180 throws java.io.IOException { 2181 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2182 } 2183 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2184 com.google.protobuf.CodedInputStream input) 2185 throws java.io.IOException { 2186 return PARSER.parseFrom(input); 2187 } 2188 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2189 com.google.protobuf.CodedInputStream input, 2190 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2191 throws java.io.IOException { 2192 return PARSER.parseFrom(input, extensionRegistry); 2193 } 2194 2195 public static Builder newBuilder() { return Builder.create(); } 2196 public Builder newBuilderForType() { return newBuilder(); } 2197 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection prototype) { 2198 return newBuilder().mergeFrom(prototype); 2199 } 2200 public Builder toBuilder() { return newBuilder(this); } 2201 2202 @java.lang.Override 2203 protected Builder newBuilderForType( 2204 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2205 Builder builder = new Builder(parent); 2206 return builder; 2207 } 2208 /** 2209 * Protobuf type {@code hadoop.hdfs.fsimage.NameSystemSection} 2210 * 2211 * <pre> 2212 ** 2213 * Name: NS_INFO 2214 * </pre> 2215 */ 2216 public static final class Builder extends 2217 com.google.protobuf.GeneratedMessage.Builder<Builder> 2218 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSectionOrBuilder { 2219 public static final com.google.protobuf.Descriptors.Descriptor 2220 getDescriptor() { 2221 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 2222 } 2223 2224 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2225 internalGetFieldAccessorTable() { 2226 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable 2227 .ensureFieldAccessorsInitialized( 2228 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.Builder.class); 2229 } 2230 2231 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.newBuilder() 2232 private Builder() { 2233 maybeForceBuilderInitialization(); 2234 } 2235 2236 private Builder( 2237 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2238 super(parent); 2239 maybeForceBuilderInitialization(); 2240 } 2241 private void maybeForceBuilderInitialization() { 2242 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2243 } 2244 } 2245 private static Builder create() { 2246 return new Builder(); 2247 } 2248 2249 public Builder clear() { 2250 super.clear(); 2251 namespaceId_ = 0; 2252 bitField0_ = (bitField0_ & ~0x00000001); 2253 genstampV1_ = 0L; 2254 bitField0_ = (bitField0_ & ~0x00000002); 2255 genstampV2_ = 0L; 2256 bitField0_ = (bitField0_ & ~0x00000004); 2257 genstampV1Limit_ = 0L; 2258 bitField0_ = (bitField0_ & ~0x00000008); 2259 lastAllocatedBlockId_ = 0L; 2260 bitField0_ = (bitField0_ & ~0x00000010); 2261 transactionId_ = 0L; 2262 bitField0_ = (bitField0_ & ~0x00000020); 2263 rollingUpgradeStartTime_ = 0L; 2264 bitField0_ = (bitField0_ & ~0x00000040); 2265 return this; 2266 } 2267 2268 public Builder clone() { 2269 return create().mergeFrom(buildPartial()); 2270 } 2271 2272 public com.google.protobuf.Descriptors.Descriptor 2273 getDescriptorForType() { 2274 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 2275 } 2276 2277 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection getDefaultInstanceForType() { 2278 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.getDefaultInstance(); 2279 } 2280 2281 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection build() { 2282 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection result = buildPartial(); 2283 if (!result.isInitialized()) { 2284 throw newUninitializedMessageException(result); 2285 } 2286 return result; 2287 } 2288 2289 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection buildPartial() { 2290 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection(this); 2291 int from_bitField0_ = bitField0_; 2292 int to_bitField0_ = 0; 2293 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 2294 to_bitField0_ |= 0x00000001; 2295 } 2296 result.namespaceId_ = namespaceId_; 2297 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 2298 to_bitField0_ |= 0x00000002; 2299 } 2300 result.genstampV1_ = genstampV1_; 2301 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 2302 to_bitField0_ |= 0x00000004; 2303 } 2304 result.genstampV2_ = genstampV2_; 2305 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 2306 to_bitField0_ |= 0x00000008; 2307 } 2308 result.genstampV1Limit_ = genstampV1Limit_; 2309 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 2310 to_bitField0_ |= 0x00000010; 2311 } 2312 result.lastAllocatedBlockId_ = lastAllocatedBlockId_; 2313 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 2314 to_bitField0_ |= 0x00000020; 2315 } 2316 result.transactionId_ = transactionId_; 2317 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 2318 to_bitField0_ |= 0x00000040; 2319 } 2320 result.rollingUpgradeStartTime_ = rollingUpgradeStartTime_; 2321 result.bitField0_ = to_bitField0_; 2322 onBuilt(); 2323 return result; 2324 } 2325 2326 public Builder mergeFrom(com.google.protobuf.Message other) { 2327 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection) { 2328 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection)other); 2329 } else { 2330 super.mergeFrom(other); 2331 return this; 2332 } 2333 } 2334 2335 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection other) { 2336 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.getDefaultInstance()) return this; 2337 if (other.hasNamespaceId()) { 2338 setNamespaceId(other.getNamespaceId()); 2339 } 2340 if (other.hasGenstampV1()) { 2341 setGenstampV1(other.getGenstampV1()); 2342 } 2343 if (other.hasGenstampV2()) { 2344 setGenstampV2(other.getGenstampV2()); 2345 } 2346 if (other.hasGenstampV1Limit()) { 2347 setGenstampV1Limit(other.getGenstampV1Limit()); 2348 } 2349 if (other.hasLastAllocatedBlockId()) { 2350 setLastAllocatedBlockId(other.getLastAllocatedBlockId()); 2351 } 2352 if (other.hasTransactionId()) { 2353 setTransactionId(other.getTransactionId()); 2354 } 2355 if (other.hasRollingUpgradeStartTime()) { 2356 setRollingUpgradeStartTime(other.getRollingUpgradeStartTime()); 2357 } 2358 this.mergeUnknownFields(other.getUnknownFields()); 2359 return this; 2360 } 2361 2362 public final boolean isInitialized() { 2363 return true; 2364 } 2365 2366 public Builder mergeFrom( 2367 com.google.protobuf.CodedInputStream input, 2368 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2369 throws java.io.IOException { 2370 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parsedMessage = null; 2371 try { 2372 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2373 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2374 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection) e.getUnfinishedMessage(); 2375 throw e; 2376 } finally { 2377 if (parsedMessage != null) { 2378 mergeFrom(parsedMessage); 2379 } 2380 } 2381 return this; 2382 } 2383 private int bitField0_; 2384 2385 // optional uint32 namespaceId = 1; 2386 private int namespaceId_ ; 2387 /** 2388 * <code>optional uint32 namespaceId = 1;</code> 2389 */ 2390 public boolean hasNamespaceId() { 2391 return ((bitField0_ & 0x00000001) == 0x00000001); 2392 } 2393 /** 2394 * <code>optional uint32 namespaceId = 1;</code> 2395 */ 2396 public int getNamespaceId() { 2397 return namespaceId_; 2398 } 2399 /** 2400 * <code>optional uint32 namespaceId = 1;</code> 2401 */ 2402 public Builder setNamespaceId(int value) { 2403 bitField0_ |= 0x00000001; 2404 namespaceId_ = value; 2405 onChanged(); 2406 return this; 2407 } 2408 /** 2409 * <code>optional uint32 namespaceId = 1;</code> 2410 */ 2411 public Builder clearNamespaceId() { 2412 bitField0_ = (bitField0_ & ~0x00000001); 2413 namespaceId_ = 0; 2414 onChanged(); 2415 return this; 2416 } 2417 2418 // optional uint64 genstampV1 = 2; 2419 private long genstampV1_ ; 2420 /** 2421 * <code>optional uint64 genstampV1 = 2;</code> 2422 */ 2423 public boolean hasGenstampV1() { 2424 return ((bitField0_ & 0x00000002) == 0x00000002); 2425 } 2426 /** 2427 * <code>optional uint64 genstampV1 = 2;</code> 2428 */ 2429 public long getGenstampV1() { 2430 return genstampV1_; 2431 } 2432 /** 2433 * <code>optional uint64 genstampV1 = 2;</code> 2434 */ 2435 public Builder setGenstampV1(long value) { 2436 bitField0_ |= 0x00000002; 2437 genstampV1_ = value; 2438 onChanged(); 2439 return this; 2440 } 2441 /** 2442 * <code>optional uint64 genstampV1 = 2;</code> 2443 */ 2444 public Builder clearGenstampV1() { 2445 bitField0_ = (bitField0_ & ~0x00000002); 2446 genstampV1_ = 0L; 2447 onChanged(); 2448 return this; 2449 } 2450 2451 // optional uint64 genstampV2 = 3; 2452 private long genstampV2_ ; 2453 /** 2454 * <code>optional uint64 genstampV2 = 3;</code> 2455 */ 2456 public boolean hasGenstampV2() { 2457 return ((bitField0_ & 0x00000004) == 0x00000004); 2458 } 2459 /** 2460 * <code>optional uint64 genstampV2 = 3;</code> 2461 */ 2462 public long getGenstampV2() { 2463 return genstampV2_; 2464 } 2465 /** 2466 * <code>optional uint64 genstampV2 = 3;</code> 2467 */ 2468 public Builder setGenstampV2(long value) { 2469 bitField0_ |= 0x00000004; 2470 genstampV2_ = value; 2471 onChanged(); 2472 return this; 2473 } 2474 /** 2475 * <code>optional uint64 genstampV2 = 3;</code> 2476 */ 2477 public Builder clearGenstampV2() { 2478 bitField0_ = (bitField0_ & ~0x00000004); 2479 genstampV2_ = 0L; 2480 onChanged(); 2481 return this; 2482 } 2483 2484 // optional uint64 genstampV1Limit = 4; 2485 private long genstampV1Limit_ ; 2486 /** 2487 * <code>optional uint64 genstampV1Limit = 4;</code> 2488 */ 2489 public boolean hasGenstampV1Limit() { 2490 return ((bitField0_ & 0x00000008) == 0x00000008); 2491 } 2492 /** 2493 * <code>optional uint64 genstampV1Limit = 4;</code> 2494 */ 2495 public long getGenstampV1Limit() { 2496 return genstampV1Limit_; 2497 } 2498 /** 2499 * <code>optional uint64 genstampV1Limit = 4;</code> 2500 */ 2501 public Builder setGenstampV1Limit(long value) { 2502 bitField0_ |= 0x00000008; 2503 genstampV1Limit_ = value; 2504 onChanged(); 2505 return this; 2506 } 2507 /** 2508 * <code>optional uint64 genstampV1Limit = 4;</code> 2509 */ 2510 public Builder clearGenstampV1Limit() { 2511 bitField0_ = (bitField0_ & ~0x00000008); 2512 genstampV1Limit_ = 0L; 2513 onChanged(); 2514 return this; 2515 } 2516 2517 // optional uint64 lastAllocatedBlockId = 5; 2518 private long lastAllocatedBlockId_ ; 2519 /** 2520 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2521 */ 2522 public boolean hasLastAllocatedBlockId() { 2523 return ((bitField0_ & 0x00000010) == 0x00000010); 2524 } 2525 /** 2526 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2527 */ 2528 public long getLastAllocatedBlockId() { 2529 return lastAllocatedBlockId_; 2530 } 2531 /** 2532 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2533 */ 2534 public Builder setLastAllocatedBlockId(long value) { 2535 bitField0_ |= 0x00000010; 2536 lastAllocatedBlockId_ = value; 2537 onChanged(); 2538 return this; 2539 } 2540 /** 2541 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2542 */ 2543 public Builder clearLastAllocatedBlockId() { 2544 bitField0_ = (bitField0_ & ~0x00000010); 2545 lastAllocatedBlockId_ = 0L; 2546 onChanged(); 2547 return this; 2548 } 2549 2550 // optional uint64 transactionId = 6; 2551 private long transactionId_ ; 2552 /** 2553 * <code>optional uint64 transactionId = 6;</code> 2554 */ 2555 public boolean hasTransactionId() { 2556 return ((bitField0_ & 0x00000020) == 0x00000020); 2557 } 2558 /** 2559 * <code>optional uint64 transactionId = 6;</code> 2560 */ 2561 public long getTransactionId() { 2562 return transactionId_; 2563 } 2564 /** 2565 * <code>optional uint64 transactionId = 6;</code> 2566 */ 2567 public Builder setTransactionId(long value) { 2568 bitField0_ |= 0x00000020; 2569 transactionId_ = value; 2570 onChanged(); 2571 return this; 2572 } 2573 /** 2574 * <code>optional uint64 transactionId = 6;</code> 2575 */ 2576 public Builder clearTransactionId() { 2577 bitField0_ = (bitField0_ & ~0x00000020); 2578 transactionId_ = 0L; 2579 onChanged(); 2580 return this; 2581 } 2582 2583 // optional uint64 rollingUpgradeStartTime = 7; 2584 private long rollingUpgradeStartTime_ ; 2585 /** 2586 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2587 */ 2588 public boolean hasRollingUpgradeStartTime() { 2589 return ((bitField0_ & 0x00000040) == 0x00000040); 2590 } 2591 /** 2592 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2593 */ 2594 public long getRollingUpgradeStartTime() { 2595 return rollingUpgradeStartTime_; 2596 } 2597 /** 2598 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2599 */ 2600 public Builder setRollingUpgradeStartTime(long value) { 2601 bitField0_ |= 0x00000040; 2602 rollingUpgradeStartTime_ = value; 2603 onChanged(); 2604 return this; 2605 } 2606 /** 2607 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2608 */ 2609 public Builder clearRollingUpgradeStartTime() { 2610 bitField0_ = (bitField0_ & ~0x00000040); 2611 rollingUpgradeStartTime_ = 0L; 2612 onChanged(); 2613 return this; 2614 } 2615 2616 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.NameSystemSection) 2617 } 2618 2619 static { 2620 defaultInstance = new NameSystemSection(true); 2621 defaultInstance.initFields(); 2622 } 2623 2624 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.NameSystemSection) 2625 } 2626 2627 public interface INodeSectionOrBuilder 2628 extends com.google.protobuf.MessageOrBuilder { 2629 2630 // optional uint64 lastInodeId = 1; 2631 /** 2632 * <code>optional uint64 lastInodeId = 1;</code> 2633 */ 2634 boolean hasLastInodeId(); 2635 /** 2636 * <code>optional uint64 lastInodeId = 1;</code> 2637 */ 2638 long getLastInodeId(); 2639 2640 // optional uint64 numInodes = 2; 2641 /** 2642 * <code>optional uint64 numInodes = 2;</code> 2643 * 2644 * <pre> 2645 * repeated INodes.. 2646 * </pre> 2647 */ 2648 boolean hasNumInodes(); 2649 /** 2650 * <code>optional uint64 numInodes = 2;</code> 2651 * 2652 * <pre> 2653 * repeated INodes.. 2654 * </pre> 2655 */ 2656 long getNumInodes(); 2657 } 2658 /** 2659 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection} 2660 * 2661 * <pre> 2662 ** 2663 * Permission is serialized as a 64-bit long. [0:24):[25:48):[48:64) (in Big Endian). 2664 * The first and the second parts are the string ids of the user and 2665 * group name, and the last 16 bits are the permission bits. 2666 * 2667 * Name: INODE 2668 * </pre> 2669 */ 2670 public static final class INodeSection extends 2671 com.google.protobuf.GeneratedMessage 2672 implements INodeSectionOrBuilder { 2673 // Use INodeSection.newBuilder() to construct. 2674 private INodeSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2675 super(builder); 2676 this.unknownFields = builder.getUnknownFields(); 2677 } 2678 private INodeSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2679 2680 private static final INodeSection defaultInstance; 2681 public static INodeSection getDefaultInstance() { 2682 return defaultInstance; 2683 } 2684 2685 public INodeSection getDefaultInstanceForType() { 2686 return defaultInstance; 2687 } 2688 2689 private final com.google.protobuf.UnknownFieldSet unknownFields; 2690 @java.lang.Override 2691 public final com.google.protobuf.UnknownFieldSet 2692 getUnknownFields() { 2693 return this.unknownFields; 2694 } 2695 private INodeSection( 2696 com.google.protobuf.CodedInputStream input, 2697 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2698 throws com.google.protobuf.InvalidProtocolBufferException { 2699 initFields(); 2700 int mutable_bitField0_ = 0; 2701 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2702 com.google.protobuf.UnknownFieldSet.newBuilder(); 2703 try { 2704 boolean done = false; 2705 while (!done) { 2706 int tag = input.readTag(); 2707 switch (tag) { 2708 case 0: 2709 done = true; 2710 break; 2711 default: { 2712 if (!parseUnknownField(input, unknownFields, 2713 extensionRegistry, tag)) { 2714 done = true; 2715 } 2716 break; 2717 } 2718 case 8: { 2719 bitField0_ |= 0x00000001; 2720 lastInodeId_ = input.readUInt64(); 2721 break; 2722 } 2723 case 16: { 2724 bitField0_ |= 0x00000002; 2725 numInodes_ = input.readUInt64(); 2726 break; 2727 } 2728 } 2729 } 2730 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2731 throw e.setUnfinishedMessage(this); 2732 } catch (java.io.IOException e) { 2733 throw new com.google.protobuf.InvalidProtocolBufferException( 2734 e.getMessage()).setUnfinishedMessage(this); 2735 } finally { 2736 this.unknownFields = unknownFields.build(); 2737 makeExtensionsImmutable(); 2738 } 2739 } 2740 public static final com.google.protobuf.Descriptors.Descriptor 2741 getDescriptor() { 2742 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 2743 } 2744 2745 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2746 internalGetFieldAccessorTable() { 2747 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable 2748 .ensureFieldAccessorsInitialized( 2749 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.Builder.class); 2750 } 2751 2752 public static com.google.protobuf.Parser<INodeSection> PARSER = 2753 new com.google.protobuf.AbstractParser<INodeSection>() { 2754 public INodeSection parsePartialFrom( 2755 com.google.protobuf.CodedInputStream input, 2756 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2757 throws com.google.protobuf.InvalidProtocolBufferException { 2758 return new INodeSection(input, extensionRegistry); 2759 } 2760 }; 2761 2762 @java.lang.Override 2763 public com.google.protobuf.Parser<INodeSection> getParserForType() { 2764 return PARSER; 2765 } 2766 2767 public interface FileUnderConstructionFeatureOrBuilder 2768 extends com.google.protobuf.MessageOrBuilder { 2769 2770 // optional string clientName = 1; 2771 /** 2772 * <code>optional string clientName = 1;</code> 2773 */ 2774 boolean hasClientName(); 2775 /** 2776 * <code>optional string clientName = 1;</code> 2777 */ 2778 java.lang.String getClientName(); 2779 /** 2780 * <code>optional string clientName = 1;</code> 2781 */ 2782 com.google.protobuf.ByteString 2783 getClientNameBytes(); 2784 2785 // optional string clientMachine = 2; 2786 /** 2787 * <code>optional string clientMachine = 2;</code> 2788 */ 2789 boolean hasClientMachine(); 2790 /** 2791 * <code>optional string clientMachine = 2;</code> 2792 */ 2793 java.lang.String getClientMachine(); 2794 /** 2795 * <code>optional string clientMachine = 2;</code> 2796 */ 2797 com.google.protobuf.ByteString 2798 getClientMachineBytes(); 2799 } 2800 /** 2801 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature} 2802 * 2803 * <pre> 2804 ** 2805 * under-construction feature for INodeFile 2806 * </pre> 2807 */ 2808 public static final class FileUnderConstructionFeature extends 2809 com.google.protobuf.GeneratedMessage 2810 implements FileUnderConstructionFeatureOrBuilder { 2811 // Use FileUnderConstructionFeature.newBuilder() to construct. 2812 private FileUnderConstructionFeature(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2813 super(builder); 2814 this.unknownFields = builder.getUnknownFields(); 2815 } 2816 private FileUnderConstructionFeature(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2817 2818 private static final FileUnderConstructionFeature defaultInstance; 2819 public static FileUnderConstructionFeature getDefaultInstance() { 2820 return defaultInstance; 2821 } 2822 2823 public FileUnderConstructionFeature getDefaultInstanceForType() { 2824 return defaultInstance; 2825 } 2826 2827 private final com.google.protobuf.UnknownFieldSet unknownFields; 2828 @java.lang.Override 2829 public final com.google.protobuf.UnknownFieldSet 2830 getUnknownFields() { 2831 return this.unknownFields; 2832 } 2833 private FileUnderConstructionFeature( 2834 com.google.protobuf.CodedInputStream input, 2835 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2836 throws com.google.protobuf.InvalidProtocolBufferException { 2837 initFields(); 2838 int mutable_bitField0_ = 0; 2839 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2840 com.google.protobuf.UnknownFieldSet.newBuilder(); 2841 try { 2842 boolean done = false; 2843 while (!done) { 2844 int tag = input.readTag(); 2845 switch (tag) { 2846 case 0: 2847 done = true; 2848 break; 2849 default: { 2850 if (!parseUnknownField(input, unknownFields, 2851 extensionRegistry, tag)) { 2852 done = true; 2853 } 2854 break; 2855 } 2856 case 10: { 2857 bitField0_ |= 0x00000001; 2858 clientName_ = input.readBytes(); 2859 break; 2860 } 2861 case 18: { 2862 bitField0_ |= 0x00000002; 2863 clientMachine_ = input.readBytes(); 2864 break; 2865 } 2866 } 2867 } 2868 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2869 throw e.setUnfinishedMessage(this); 2870 } catch (java.io.IOException e) { 2871 throw new com.google.protobuf.InvalidProtocolBufferException( 2872 e.getMessage()).setUnfinishedMessage(this); 2873 } finally { 2874 this.unknownFields = unknownFields.build(); 2875 makeExtensionsImmutable(); 2876 } 2877 } 2878 public static final com.google.protobuf.Descriptors.Descriptor 2879 getDescriptor() { 2880 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 2881 } 2882 2883 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2884 internalGetFieldAccessorTable() { 2885 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable 2886 .ensureFieldAccessorsInitialized( 2887 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder.class); 2888 } 2889 2890 public static com.google.protobuf.Parser<FileUnderConstructionFeature> PARSER = 2891 new com.google.protobuf.AbstractParser<FileUnderConstructionFeature>() { 2892 public FileUnderConstructionFeature parsePartialFrom( 2893 com.google.protobuf.CodedInputStream input, 2894 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2895 throws com.google.protobuf.InvalidProtocolBufferException { 2896 return new FileUnderConstructionFeature(input, extensionRegistry); 2897 } 2898 }; 2899 2900 @java.lang.Override 2901 public com.google.protobuf.Parser<FileUnderConstructionFeature> getParserForType() { 2902 return PARSER; 2903 } 2904 2905 private int bitField0_; 2906 // optional string clientName = 1; 2907 public static final int CLIENTNAME_FIELD_NUMBER = 1; 2908 private java.lang.Object clientName_; 2909 /** 2910 * <code>optional string clientName = 1;</code> 2911 */ 2912 public boolean hasClientName() { 2913 return ((bitField0_ & 0x00000001) == 0x00000001); 2914 } 2915 /** 2916 * <code>optional string clientName = 1;</code> 2917 */ 2918 public java.lang.String getClientName() { 2919 java.lang.Object ref = clientName_; 2920 if (ref instanceof java.lang.String) { 2921 return (java.lang.String) ref; 2922 } else { 2923 com.google.protobuf.ByteString bs = 2924 (com.google.protobuf.ByteString) ref; 2925 java.lang.String s = bs.toStringUtf8(); 2926 if (bs.isValidUtf8()) { 2927 clientName_ = s; 2928 } 2929 return s; 2930 } 2931 } 2932 /** 2933 * <code>optional string clientName = 1;</code> 2934 */ 2935 public com.google.protobuf.ByteString 2936 getClientNameBytes() { 2937 java.lang.Object ref = clientName_; 2938 if (ref instanceof java.lang.String) { 2939 com.google.protobuf.ByteString b = 2940 com.google.protobuf.ByteString.copyFromUtf8( 2941 (java.lang.String) ref); 2942 clientName_ = b; 2943 return b; 2944 } else { 2945 return (com.google.protobuf.ByteString) ref; 2946 } 2947 } 2948 2949 // optional string clientMachine = 2; 2950 public static final int CLIENTMACHINE_FIELD_NUMBER = 2; 2951 private java.lang.Object clientMachine_; 2952 /** 2953 * <code>optional string clientMachine = 2;</code> 2954 */ 2955 public boolean hasClientMachine() { 2956 return ((bitField0_ & 0x00000002) == 0x00000002); 2957 } 2958 /** 2959 * <code>optional string clientMachine = 2;</code> 2960 */ 2961 public java.lang.String getClientMachine() { 2962 java.lang.Object ref = clientMachine_; 2963 if (ref instanceof java.lang.String) { 2964 return (java.lang.String) ref; 2965 } else { 2966 com.google.protobuf.ByteString bs = 2967 (com.google.protobuf.ByteString) ref; 2968 java.lang.String s = bs.toStringUtf8(); 2969 if (bs.isValidUtf8()) { 2970 clientMachine_ = s; 2971 } 2972 return s; 2973 } 2974 } 2975 /** 2976 * <code>optional string clientMachine = 2;</code> 2977 */ 2978 public com.google.protobuf.ByteString 2979 getClientMachineBytes() { 2980 java.lang.Object ref = clientMachine_; 2981 if (ref instanceof java.lang.String) { 2982 com.google.protobuf.ByteString b = 2983 com.google.protobuf.ByteString.copyFromUtf8( 2984 (java.lang.String) ref); 2985 clientMachine_ = b; 2986 return b; 2987 } else { 2988 return (com.google.protobuf.ByteString) ref; 2989 } 2990 } 2991 2992 private void initFields() { 2993 clientName_ = ""; 2994 clientMachine_ = ""; 2995 } 2996 private byte memoizedIsInitialized = -1; 2997 public final boolean isInitialized() { 2998 byte isInitialized = memoizedIsInitialized; 2999 if (isInitialized != -1) return isInitialized == 1; 3000 3001 memoizedIsInitialized = 1; 3002 return true; 3003 } 3004 3005 public void writeTo(com.google.protobuf.CodedOutputStream output) 3006 throws java.io.IOException { 3007 getSerializedSize(); 3008 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3009 output.writeBytes(1, getClientNameBytes()); 3010 } 3011 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3012 output.writeBytes(2, getClientMachineBytes()); 3013 } 3014 getUnknownFields().writeTo(output); 3015 } 3016 3017 private int memoizedSerializedSize = -1; 3018 public int getSerializedSize() { 3019 int size = memoizedSerializedSize; 3020 if (size != -1) return size; 3021 3022 size = 0; 3023 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3024 size += com.google.protobuf.CodedOutputStream 3025 .computeBytesSize(1, getClientNameBytes()); 3026 } 3027 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3028 size += com.google.protobuf.CodedOutputStream 3029 .computeBytesSize(2, getClientMachineBytes()); 3030 } 3031 size += getUnknownFields().getSerializedSize(); 3032 memoizedSerializedSize = size; 3033 return size; 3034 } 3035 3036 private static final long serialVersionUID = 0L; 3037 @java.lang.Override 3038 protected java.lang.Object writeReplace() 3039 throws java.io.ObjectStreamException { 3040 return super.writeReplace(); 3041 } 3042 3043 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3044 com.google.protobuf.ByteString data) 3045 throws com.google.protobuf.InvalidProtocolBufferException { 3046 return PARSER.parseFrom(data); 3047 } 3048 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3049 com.google.protobuf.ByteString data, 3050 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3051 throws com.google.protobuf.InvalidProtocolBufferException { 3052 return PARSER.parseFrom(data, extensionRegistry); 3053 } 3054 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom(byte[] data) 3055 throws com.google.protobuf.InvalidProtocolBufferException { 3056 return PARSER.parseFrom(data); 3057 } 3058 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3059 byte[] data, 3060 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3061 throws com.google.protobuf.InvalidProtocolBufferException { 3062 return PARSER.parseFrom(data, extensionRegistry); 3063 } 3064 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom(java.io.InputStream input) 3065 throws java.io.IOException { 3066 return PARSER.parseFrom(input); 3067 } 3068 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3069 java.io.InputStream input, 3070 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3071 throws java.io.IOException { 3072 return PARSER.parseFrom(input, extensionRegistry); 3073 } 3074 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseDelimitedFrom(java.io.InputStream input) 3075 throws java.io.IOException { 3076 return PARSER.parseDelimitedFrom(input); 3077 } 3078 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseDelimitedFrom( 3079 java.io.InputStream input, 3080 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3081 throws java.io.IOException { 3082 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3083 } 3084 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3085 com.google.protobuf.CodedInputStream input) 3086 throws java.io.IOException { 3087 return PARSER.parseFrom(input); 3088 } 3089 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3090 com.google.protobuf.CodedInputStream input, 3091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3092 throws java.io.IOException { 3093 return PARSER.parseFrom(input, extensionRegistry); 3094 } 3095 3096 public static Builder newBuilder() { return Builder.create(); } 3097 public Builder newBuilderForType() { return newBuilder(); } 3098 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature prototype) { 3099 return newBuilder().mergeFrom(prototype); 3100 } 3101 public Builder toBuilder() { return newBuilder(this); } 3102 3103 @java.lang.Override 3104 protected Builder newBuilderForType( 3105 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3106 Builder builder = new Builder(parent); 3107 return builder; 3108 } 3109 /** 3110 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature} 3111 * 3112 * <pre> 3113 ** 3114 * under-construction feature for INodeFile 3115 * </pre> 3116 */ 3117 public static final class Builder extends 3118 com.google.protobuf.GeneratedMessage.Builder<Builder> 3119 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder { 3120 public static final com.google.protobuf.Descriptors.Descriptor 3121 getDescriptor() { 3122 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 3123 } 3124 3125 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3126 internalGetFieldAccessorTable() { 3127 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable 3128 .ensureFieldAccessorsInitialized( 3129 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder.class); 3130 } 3131 3132 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.newBuilder() 3133 private Builder() { 3134 maybeForceBuilderInitialization(); 3135 } 3136 3137 private Builder( 3138 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3139 super(parent); 3140 maybeForceBuilderInitialization(); 3141 } 3142 private void maybeForceBuilderInitialization() { 3143 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3144 } 3145 } 3146 private static Builder create() { 3147 return new Builder(); 3148 } 3149 3150 public Builder clear() { 3151 super.clear(); 3152 clientName_ = ""; 3153 bitField0_ = (bitField0_ & ~0x00000001); 3154 clientMachine_ = ""; 3155 bitField0_ = (bitField0_ & ~0x00000002); 3156 return this; 3157 } 3158 3159 public Builder clone() { 3160 return create().mergeFrom(buildPartial()); 3161 } 3162 3163 public com.google.protobuf.Descriptors.Descriptor 3164 getDescriptorForType() { 3165 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 3166 } 3167 3168 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getDefaultInstanceForType() { 3169 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 3170 } 3171 3172 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature build() { 3173 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature result = buildPartial(); 3174 if (!result.isInitialized()) { 3175 throw newUninitializedMessageException(result); 3176 } 3177 return result; 3178 } 3179 3180 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature buildPartial() { 3181 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature(this); 3182 int from_bitField0_ = bitField0_; 3183 int to_bitField0_ = 0; 3184 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 3185 to_bitField0_ |= 0x00000001; 3186 } 3187 result.clientName_ = clientName_; 3188 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 3189 to_bitField0_ |= 0x00000002; 3190 } 3191 result.clientMachine_ = clientMachine_; 3192 result.bitField0_ = to_bitField0_; 3193 onBuilt(); 3194 return result; 3195 } 3196 3197 public Builder mergeFrom(com.google.protobuf.Message other) { 3198 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature) { 3199 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature)other); 3200 } else { 3201 super.mergeFrom(other); 3202 return this; 3203 } 3204 } 3205 3206 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature other) { 3207 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance()) return this; 3208 if (other.hasClientName()) { 3209 bitField0_ |= 0x00000001; 3210 clientName_ = other.clientName_; 3211 onChanged(); 3212 } 3213 if (other.hasClientMachine()) { 3214 bitField0_ |= 0x00000002; 3215 clientMachine_ = other.clientMachine_; 3216 onChanged(); 3217 } 3218 this.mergeUnknownFields(other.getUnknownFields()); 3219 return this; 3220 } 3221 3222 public final boolean isInitialized() { 3223 return true; 3224 } 3225 3226 public Builder mergeFrom( 3227 com.google.protobuf.CodedInputStream input, 3228 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3229 throws java.io.IOException { 3230 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parsedMessage = null; 3231 try { 3232 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3233 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3234 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature) e.getUnfinishedMessage(); 3235 throw e; 3236 } finally { 3237 if (parsedMessage != null) { 3238 mergeFrom(parsedMessage); 3239 } 3240 } 3241 return this; 3242 } 3243 private int bitField0_; 3244 3245 // optional string clientName = 1; 3246 private java.lang.Object clientName_ = ""; 3247 /** 3248 * <code>optional string clientName = 1;</code> 3249 */ 3250 public boolean hasClientName() { 3251 return ((bitField0_ & 0x00000001) == 0x00000001); 3252 } 3253 /** 3254 * <code>optional string clientName = 1;</code> 3255 */ 3256 public java.lang.String getClientName() { 3257 java.lang.Object ref = clientName_; 3258 if (!(ref instanceof java.lang.String)) { 3259 java.lang.String s = ((com.google.protobuf.ByteString) ref) 3260 .toStringUtf8(); 3261 clientName_ = s; 3262 return s; 3263 } else { 3264 return (java.lang.String) ref; 3265 } 3266 } 3267 /** 3268 * <code>optional string clientName = 1;</code> 3269 */ 3270 public com.google.protobuf.ByteString 3271 getClientNameBytes() { 3272 java.lang.Object ref = clientName_; 3273 if (ref instanceof String) { 3274 com.google.protobuf.ByteString b = 3275 com.google.protobuf.ByteString.copyFromUtf8( 3276 (java.lang.String) ref); 3277 clientName_ = b; 3278 return b; 3279 } else { 3280 return (com.google.protobuf.ByteString) ref; 3281 } 3282 } 3283 /** 3284 * <code>optional string clientName = 1;</code> 3285 */ 3286 public Builder setClientName( 3287 java.lang.String value) { 3288 if (value == null) { 3289 throw new NullPointerException(); 3290 } 3291 bitField0_ |= 0x00000001; 3292 clientName_ = value; 3293 onChanged(); 3294 return this; 3295 } 3296 /** 3297 * <code>optional string clientName = 1;</code> 3298 */ 3299 public Builder clearClientName() { 3300 bitField0_ = (bitField0_ & ~0x00000001); 3301 clientName_ = getDefaultInstance().getClientName(); 3302 onChanged(); 3303 return this; 3304 } 3305 /** 3306 * <code>optional string clientName = 1;</code> 3307 */ 3308 public Builder setClientNameBytes( 3309 com.google.protobuf.ByteString value) { 3310 if (value == null) { 3311 throw new NullPointerException(); 3312 } 3313 bitField0_ |= 0x00000001; 3314 clientName_ = value; 3315 onChanged(); 3316 return this; 3317 } 3318 3319 // optional string clientMachine = 2; 3320 private java.lang.Object clientMachine_ = ""; 3321 /** 3322 * <code>optional string clientMachine = 2;</code> 3323 */ 3324 public boolean hasClientMachine() { 3325 return ((bitField0_ & 0x00000002) == 0x00000002); 3326 } 3327 /** 3328 * <code>optional string clientMachine = 2;</code> 3329 */ 3330 public java.lang.String getClientMachine() { 3331 java.lang.Object ref = clientMachine_; 3332 if (!(ref instanceof java.lang.String)) { 3333 java.lang.String s = ((com.google.protobuf.ByteString) ref) 3334 .toStringUtf8(); 3335 clientMachine_ = s; 3336 return s; 3337 } else { 3338 return (java.lang.String) ref; 3339 } 3340 } 3341 /** 3342 * <code>optional string clientMachine = 2;</code> 3343 */ 3344 public com.google.protobuf.ByteString 3345 getClientMachineBytes() { 3346 java.lang.Object ref = clientMachine_; 3347 if (ref instanceof String) { 3348 com.google.protobuf.ByteString b = 3349 com.google.protobuf.ByteString.copyFromUtf8( 3350 (java.lang.String) ref); 3351 clientMachine_ = b; 3352 return b; 3353 } else { 3354 return (com.google.protobuf.ByteString) ref; 3355 } 3356 } 3357 /** 3358 * <code>optional string clientMachine = 2;</code> 3359 */ 3360 public Builder setClientMachine( 3361 java.lang.String value) { 3362 if (value == null) { 3363 throw new NullPointerException(); 3364 } 3365 bitField0_ |= 0x00000002; 3366 clientMachine_ = value; 3367 onChanged(); 3368 return this; 3369 } 3370 /** 3371 * <code>optional string clientMachine = 2;</code> 3372 */ 3373 public Builder clearClientMachine() { 3374 bitField0_ = (bitField0_ & ~0x00000002); 3375 clientMachine_ = getDefaultInstance().getClientMachine(); 3376 onChanged(); 3377 return this; 3378 } 3379 /** 3380 * <code>optional string clientMachine = 2;</code> 3381 */ 3382 public Builder setClientMachineBytes( 3383 com.google.protobuf.ByteString value) { 3384 if (value == null) { 3385 throw new NullPointerException(); 3386 } 3387 bitField0_ |= 0x00000002; 3388 clientMachine_ = value; 3389 onChanged(); 3390 return this; 3391 } 3392 3393 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature) 3394 } 3395 3396 static { 3397 defaultInstance = new FileUnderConstructionFeature(true); 3398 defaultInstance.initFields(); 3399 } 3400 3401 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature) 3402 } 3403 3404 public interface AclFeatureProtoOrBuilder 3405 extends com.google.protobuf.MessageOrBuilder { 3406 3407 // repeated fixed32 entries = 2 [packed = true]; 3408 /** 3409 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3410 * 3411 * <pre> 3412 ** 3413 * An ACL entry is represented by a 32-bit integer in Big Endian 3414 * format. The bits can be divided in four segments: 3415 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3416 * 3417 * [0:2) -- reserved for futute uses. 3418 * [2:26) -- the name of the entry, which is an ID that points to a 3419 * string in the StringTableSection. 3420 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3421 * [27:29) -- the type of the entry (AclEntryTypeProto) 3422 * [29:32) -- the permission of the entry (FsActionProto) 3423 * </pre> 3424 */ 3425 java.util.List<java.lang.Integer> getEntriesList(); 3426 /** 3427 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3428 * 3429 * <pre> 3430 ** 3431 * An ACL entry is represented by a 32-bit integer in Big Endian 3432 * format. The bits can be divided in four segments: 3433 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3434 * 3435 * [0:2) -- reserved for futute uses. 3436 * [2:26) -- the name of the entry, which is an ID that points to a 3437 * string in the StringTableSection. 3438 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3439 * [27:29) -- the type of the entry (AclEntryTypeProto) 3440 * [29:32) -- the permission of the entry (FsActionProto) 3441 * </pre> 3442 */ 3443 int getEntriesCount(); 3444 /** 3445 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3446 * 3447 * <pre> 3448 ** 3449 * An ACL entry is represented by a 32-bit integer in Big Endian 3450 * format. The bits can be divided in four segments: 3451 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3452 * 3453 * [0:2) -- reserved for futute uses. 3454 * [2:26) -- the name of the entry, which is an ID that points to a 3455 * string in the StringTableSection. 3456 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3457 * [27:29) -- the type of the entry (AclEntryTypeProto) 3458 * [29:32) -- the permission of the entry (FsActionProto) 3459 * </pre> 3460 */ 3461 int getEntries(int index); 3462 } 3463 /** 3464 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.AclFeatureProto} 3465 */ 3466 public static final class AclFeatureProto extends 3467 com.google.protobuf.GeneratedMessage 3468 implements AclFeatureProtoOrBuilder { 3469 // Use AclFeatureProto.newBuilder() to construct. 3470 private AclFeatureProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3471 super(builder); 3472 this.unknownFields = builder.getUnknownFields(); 3473 } 3474 private AclFeatureProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3475 3476 private static final AclFeatureProto defaultInstance; 3477 public static AclFeatureProto getDefaultInstance() { 3478 return defaultInstance; 3479 } 3480 3481 public AclFeatureProto getDefaultInstanceForType() { 3482 return defaultInstance; 3483 } 3484 3485 private final com.google.protobuf.UnknownFieldSet unknownFields; 3486 @java.lang.Override 3487 public final com.google.protobuf.UnknownFieldSet 3488 getUnknownFields() { 3489 return this.unknownFields; 3490 } 3491 private AclFeatureProto( 3492 com.google.protobuf.CodedInputStream input, 3493 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3494 throws com.google.protobuf.InvalidProtocolBufferException { 3495 initFields(); 3496 int mutable_bitField0_ = 0; 3497 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3498 com.google.protobuf.UnknownFieldSet.newBuilder(); 3499 try { 3500 boolean done = false; 3501 while (!done) { 3502 int tag = input.readTag(); 3503 switch (tag) { 3504 case 0: 3505 done = true; 3506 break; 3507 default: { 3508 if (!parseUnknownField(input, unknownFields, 3509 extensionRegistry, tag)) { 3510 done = true; 3511 } 3512 break; 3513 } 3514 case 21: { 3515 if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 3516 entries_ = new java.util.ArrayList<java.lang.Integer>(); 3517 mutable_bitField0_ |= 0x00000001; 3518 } 3519 entries_.add(input.readFixed32()); 3520 break; 3521 } 3522 case 18: { 3523 int length = input.readRawVarint32(); 3524 int limit = input.pushLimit(length); 3525 if (!((mutable_bitField0_ & 0x00000001) == 0x00000001) && input.getBytesUntilLimit() > 0) { 3526 entries_ = new java.util.ArrayList<java.lang.Integer>(); 3527 mutable_bitField0_ |= 0x00000001; 3528 } 3529 while (input.getBytesUntilLimit() > 0) { 3530 entries_.add(input.readFixed32()); 3531 } 3532 input.popLimit(limit); 3533 break; 3534 } 3535 } 3536 } 3537 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3538 throw e.setUnfinishedMessage(this); 3539 } catch (java.io.IOException e) { 3540 throw new com.google.protobuf.InvalidProtocolBufferException( 3541 e.getMessage()).setUnfinishedMessage(this); 3542 } finally { 3543 if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 3544 entries_ = java.util.Collections.unmodifiableList(entries_); 3545 } 3546 this.unknownFields = unknownFields.build(); 3547 makeExtensionsImmutable(); 3548 } 3549 } 3550 public static final com.google.protobuf.Descriptors.Descriptor 3551 getDescriptor() { 3552 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3553 } 3554 3555 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3556 internalGetFieldAccessorTable() { 3557 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable 3558 .ensureFieldAccessorsInitialized( 3559 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder.class); 3560 } 3561 3562 public static com.google.protobuf.Parser<AclFeatureProto> PARSER = 3563 new com.google.protobuf.AbstractParser<AclFeatureProto>() { 3564 public AclFeatureProto parsePartialFrom( 3565 com.google.protobuf.CodedInputStream input, 3566 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3567 throws com.google.protobuf.InvalidProtocolBufferException { 3568 return new AclFeatureProto(input, extensionRegistry); 3569 } 3570 }; 3571 3572 @java.lang.Override 3573 public com.google.protobuf.Parser<AclFeatureProto> getParserForType() { 3574 return PARSER; 3575 } 3576 3577 // repeated fixed32 entries = 2 [packed = true]; 3578 public static final int ENTRIES_FIELD_NUMBER = 2; 3579 private java.util.List<java.lang.Integer> entries_; 3580 /** 3581 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3582 * 3583 * <pre> 3584 ** 3585 * An ACL entry is represented by a 32-bit integer in Big Endian 3586 * format. The bits can be divided in four segments: 3587 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3588 * 3589 * [0:2) -- reserved for futute uses. 3590 * [2:26) -- the name of the entry, which is an ID that points to a 3591 * string in the StringTableSection. 3592 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3593 * [27:29) -- the type of the entry (AclEntryTypeProto) 3594 * [29:32) -- the permission of the entry (FsActionProto) 3595 * </pre> 3596 */ 3597 public java.util.List<java.lang.Integer> 3598 getEntriesList() { 3599 return entries_; 3600 } 3601 /** 3602 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3603 * 3604 * <pre> 3605 ** 3606 * An ACL entry is represented by a 32-bit integer in Big Endian 3607 * format. The bits can be divided in four segments: 3608 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3609 * 3610 * [0:2) -- reserved for futute uses. 3611 * [2:26) -- the name of the entry, which is an ID that points to a 3612 * string in the StringTableSection. 3613 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3614 * [27:29) -- the type of the entry (AclEntryTypeProto) 3615 * [29:32) -- the permission of the entry (FsActionProto) 3616 * </pre> 3617 */ 3618 public int getEntriesCount() { 3619 return entries_.size(); 3620 } 3621 /** 3622 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3623 * 3624 * <pre> 3625 ** 3626 * An ACL entry is represented by a 32-bit integer in Big Endian 3627 * format. The bits can be divided in four segments: 3628 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3629 * 3630 * [0:2) -- reserved for futute uses. 3631 * [2:26) -- the name of the entry, which is an ID that points to a 3632 * string in the StringTableSection. 3633 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3634 * [27:29) -- the type of the entry (AclEntryTypeProto) 3635 * [29:32) -- the permission of the entry (FsActionProto) 3636 * </pre> 3637 */ 3638 public int getEntries(int index) { 3639 return entries_.get(index); 3640 } 3641 private int entriesMemoizedSerializedSize = -1; 3642 3643 private void initFields() { 3644 entries_ = java.util.Collections.emptyList(); 3645 } 3646 private byte memoizedIsInitialized = -1; 3647 public final boolean isInitialized() { 3648 byte isInitialized = memoizedIsInitialized; 3649 if (isInitialized != -1) return isInitialized == 1; 3650 3651 memoizedIsInitialized = 1; 3652 return true; 3653 } 3654 3655 public void writeTo(com.google.protobuf.CodedOutputStream output) 3656 throws java.io.IOException { 3657 getSerializedSize(); 3658 if (getEntriesList().size() > 0) { 3659 output.writeRawVarint32(18); 3660 output.writeRawVarint32(entriesMemoizedSerializedSize); 3661 } 3662 for (int i = 0; i < entries_.size(); i++) { 3663 output.writeFixed32NoTag(entries_.get(i)); 3664 } 3665 getUnknownFields().writeTo(output); 3666 } 3667 3668 private int memoizedSerializedSize = -1; 3669 public int getSerializedSize() { 3670 int size = memoizedSerializedSize; 3671 if (size != -1) return size; 3672 3673 size = 0; 3674 { 3675 int dataSize = 0; 3676 dataSize = 4 * getEntriesList().size(); 3677 size += dataSize; 3678 if (!getEntriesList().isEmpty()) { 3679 size += 1; 3680 size += com.google.protobuf.CodedOutputStream 3681 .computeInt32SizeNoTag(dataSize); 3682 } 3683 entriesMemoizedSerializedSize = dataSize; 3684 } 3685 size += getUnknownFields().getSerializedSize(); 3686 memoizedSerializedSize = size; 3687 return size; 3688 } 3689 3690 private static final long serialVersionUID = 0L; 3691 @java.lang.Override 3692 protected java.lang.Object writeReplace() 3693 throws java.io.ObjectStreamException { 3694 return super.writeReplace(); 3695 } 3696 3697 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3698 com.google.protobuf.ByteString data) 3699 throws com.google.protobuf.InvalidProtocolBufferException { 3700 return PARSER.parseFrom(data); 3701 } 3702 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3703 com.google.protobuf.ByteString data, 3704 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3705 throws com.google.protobuf.InvalidProtocolBufferException { 3706 return PARSER.parseFrom(data, extensionRegistry); 3707 } 3708 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom(byte[] data) 3709 throws com.google.protobuf.InvalidProtocolBufferException { 3710 return PARSER.parseFrom(data); 3711 } 3712 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3713 byte[] data, 3714 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3715 throws com.google.protobuf.InvalidProtocolBufferException { 3716 return PARSER.parseFrom(data, extensionRegistry); 3717 } 3718 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom(java.io.InputStream input) 3719 throws java.io.IOException { 3720 return PARSER.parseFrom(input); 3721 } 3722 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3723 java.io.InputStream input, 3724 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3725 throws java.io.IOException { 3726 return PARSER.parseFrom(input, extensionRegistry); 3727 } 3728 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseDelimitedFrom(java.io.InputStream input) 3729 throws java.io.IOException { 3730 return PARSER.parseDelimitedFrom(input); 3731 } 3732 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseDelimitedFrom( 3733 java.io.InputStream input, 3734 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3735 throws java.io.IOException { 3736 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3737 } 3738 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3739 com.google.protobuf.CodedInputStream input) 3740 throws java.io.IOException { 3741 return PARSER.parseFrom(input); 3742 } 3743 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3744 com.google.protobuf.CodedInputStream input, 3745 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3746 throws java.io.IOException { 3747 return PARSER.parseFrom(input, extensionRegistry); 3748 } 3749 3750 public static Builder newBuilder() { return Builder.create(); } 3751 public Builder newBuilderForType() { return newBuilder(); } 3752 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto prototype) { 3753 return newBuilder().mergeFrom(prototype); 3754 } 3755 public Builder toBuilder() { return newBuilder(this); } 3756 3757 @java.lang.Override 3758 protected Builder newBuilderForType( 3759 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3760 Builder builder = new Builder(parent); 3761 return builder; 3762 } 3763 /** 3764 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.AclFeatureProto} 3765 */ 3766 public static final class Builder extends 3767 com.google.protobuf.GeneratedMessage.Builder<Builder> 3768 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder { 3769 public static final com.google.protobuf.Descriptors.Descriptor 3770 getDescriptor() { 3771 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3772 } 3773 3774 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3775 internalGetFieldAccessorTable() { 3776 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable 3777 .ensureFieldAccessorsInitialized( 3778 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder.class); 3779 } 3780 3781 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder() 3782 private Builder() { 3783 maybeForceBuilderInitialization(); 3784 } 3785 3786 private Builder( 3787 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3788 super(parent); 3789 maybeForceBuilderInitialization(); 3790 } 3791 private void maybeForceBuilderInitialization() { 3792 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3793 } 3794 } 3795 private static Builder create() { 3796 return new Builder(); 3797 } 3798 3799 public Builder clear() { 3800 super.clear(); 3801 entries_ = java.util.Collections.emptyList(); 3802 bitField0_ = (bitField0_ & ~0x00000001); 3803 return this; 3804 } 3805 3806 public Builder clone() { 3807 return create().mergeFrom(buildPartial()); 3808 } 3809 3810 public com.google.protobuf.Descriptors.Descriptor 3811 getDescriptorForType() { 3812 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3813 } 3814 3815 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getDefaultInstanceForType() { 3816 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 3817 } 3818 3819 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto build() { 3820 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto result = buildPartial(); 3821 if (!result.isInitialized()) { 3822 throw newUninitializedMessageException(result); 3823 } 3824 return result; 3825 } 3826 3827 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto buildPartial() { 3828 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto(this); 3829 int from_bitField0_ = bitField0_; 3830 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3831 entries_ = java.util.Collections.unmodifiableList(entries_); 3832 bitField0_ = (bitField0_ & ~0x00000001); 3833 } 3834 result.entries_ = entries_; 3835 onBuilt(); 3836 return result; 3837 } 3838 3839 public Builder mergeFrom(com.google.protobuf.Message other) { 3840 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto) { 3841 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto)other); 3842 } else { 3843 super.mergeFrom(other); 3844 return this; 3845 } 3846 } 3847 3848 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto other) { 3849 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) return this; 3850 if (!other.entries_.isEmpty()) { 3851 if (entries_.isEmpty()) { 3852 entries_ = other.entries_; 3853 bitField0_ = (bitField0_ & ~0x00000001); 3854 } else { 3855 ensureEntriesIsMutable(); 3856 entries_.addAll(other.entries_); 3857 } 3858 onChanged(); 3859 } 3860 this.mergeUnknownFields(other.getUnknownFields()); 3861 return this; 3862 } 3863 3864 public final boolean isInitialized() { 3865 return true; 3866 } 3867 3868 public Builder mergeFrom( 3869 com.google.protobuf.CodedInputStream input, 3870 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3871 throws java.io.IOException { 3872 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parsedMessage = null; 3873 try { 3874 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3875 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3876 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto) e.getUnfinishedMessage(); 3877 throw e; 3878 } finally { 3879 if (parsedMessage != null) { 3880 mergeFrom(parsedMessage); 3881 } 3882 } 3883 return this; 3884 } 3885 private int bitField0_; 3886 3887 // repeated fixed32 entries = 2 [packed = true]; 3888 private java.util.List<java.lang.Integer> entries_ = java.util.Collections.emptyList(); 3889 private void ensureEntriesIsMutable() { 3890 if (!((bitField0_ & 0x00000001) == 0x00000001)) { 3891 entries_ = new java.util.ArrayList<java.lang.Integer>(entries_); 3892 bitField0_ |= 0x00000001; 3893 } 3894 } 3895 /** 3896 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3897 * 3898 * <pre> 3899 ** 3900 * An ACL entry is represented by a 32-bit integer in Big Endian 3901 * format. The bits can be divided in four segments: 3902 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3903 * 3904 * [0:2) -- reserved for futute uses. 3905 * [2:26) -- the name of the entry, which is an ID that points to a 3906 * string in the StringTableSection. 3907 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3908 * [27:29) -- the type of the entry (AclEntryTypeProto) 3909 * [29:32) -- the permission of the entry (FsActionProto) 3910 * </pre> 3911 */ 3912 public java.util.List<java.lang.Integer> 3913 getEntriesList() { 3914 return java.util.Collections.unmodifiableList(entries_); 3915 } 3916 /** 3917 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3918 * 3919 * <pre> 3920 ** 3921 * An ACL entry is represented by a 32-bit integer in Big Endian 3922 * format. The bits can be divided in four segments: 3923 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3924 * 3925 * [0:2) -- reserved for futute uses. 3926 * [2:26) -- the name of the entry, which is an ID that points to a 3927 * string in the StringTableSection. 3928 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3929 * [27:29) -- the type of the entry (AclEntryTypeProto) 3930 * [29:32) -- the permission of the entry (FsActionProto) 3931 * </pre> 3932 */ 3933 public int getEntriesCount() { 3934 return entries_.size(); 3935 } 3936 /** 3937 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3938 * 3939 * <pre> 3940 ** 3941 * An ACL entry is represented by a 32-bit integer in Big Endian 3942 * format. The bits can be divided in four segments: 3943 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3944 * 3945 * [0:2) -- reserved for futute uses. 3946 * [2:26) -- the name of the entry, which is an ID that points to a 3947 * string in the StringTableSection. 3948 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3949 * [27:29) -- the type of the entry (AclEntryTypeProto) 3950 * [29:32) -- the permission of the entry (FsActionProto) 3951 * </pre> 3952 */ 3953 public int getEntries(int index) { 3954 return entries_.get(index); 3955 } 3956 /** 3957 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3958 * 3959 * <pre> 3960 ** 3961 * An ACL entry is represented by a 32-bit integer in Big Endian 3962 * format. The bits can be divided in four segments: 3963 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3964 * 3965 * [0:2) -- reserved for futute uses. 3966 * [2:26) -- the name of the entry, which is an ID that points to a 3967 * string in the StringTableSection. 3968 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3969 * [27:29) -- the type of the entry (AclEntryTypeProto) 3970 * [29:32) -- the permission of the entry (FsActionProto) 3971 * </pre> 3972 */ 3973 public Builder setEntries( 3974 int index, int value) { 3975 ensureEntriesIsMutable(); 3976 entries_.set(index, value); 3977 onChanged(); 3978 return this; 3979 } 3980 /** 3981 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3982 * 3983 * <pre> 3984 ** 3985 * An ACL entry is represented by a 32-bit integer in Big Endian 3986 * format. The bits can be divided in four segments: 3987 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3988 * 3989 * [0:2) -- reserved for futute uses. 3990 * [2:26) -- the name of the entry, which is an ID that points to a 3991 * string in the StringTableSection. 3992 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3993 * [27:29) -- the type of the entry (AclEntryTypeProto) 3994 * [29:32) -- the permission of the entry (FsActionProto) 3995 * </pre> 3996 */ 3997 public Builder addEntries(int value) { 3998 ensureEntriesIsMutable(); 3999 entries_.add(value); 4000 onChanged(); 4001 return this; 4002 } 4003 /** 4004 * <code>repeated fixed32 entries = 2 [packed = true];</code> 4005 * 4006 * <pre> 4007 ** 4008 * An ACL entry is represented by a 32-bit integer in Big Endian 4009 * format. The bits can be divided in four segments: 4010 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 4011 * 4012 * [0:2) -- reserved for futute uses. 4013 * [2:26) -- the name of the entry, which is an ID that points to a 4014 * string in the StringTableSection. 4015 * [26:27) -- the scope of the entry (AclEntryScopeProto) 4016 * [27:29) -- the type of the entry (AclEntryTypeProto) 4017 * [29:32) -- the permission of the entry (FsActionProto) 4018 * </pre> 4019 */ 4020 public Builder addAllEntries( 4021 java.lang.Iterable<? extends java.lang.Integer> values) { 4022 ensureEntriesIsMutable(); 4023 super.addAll(values, entries_); 4024 onChanged(); 4025 return this; 4026 } 4027 /** 4028 * <code>repeated fixed32 entries = 2 [packed = true];</code> 4029 * 4030 * <pre> 4031 ** 4032 * An ACL entry is represented by a 32-bit integer in Big Endian 4033 * format. The bits can be divided in four segments: 4034 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 4035 * 4036 * [0:2) -- reserved for futute uses. 4037 * [2:26) -- the name of the entry, which is an ID that points to a 4038 * string in the StringTableSection. 4039 * [26:27) -- the scope of the entry (AclEntryScopeProto) 4040 * [27:29) -- the type of the entry (AclEntryTypeProto) 4041 * [29:32) -- the permission of the entry (FsActionProto) 4042 * </pre> 4043 */ 4044 public Builder clearEntries() { 4045 entries_ = java.util.Collections.emptyList(); 4046 bitField0_ = (bitField0_ & ~0x00000001); 4047 onChanged(); 4048 return this; 4049 } 4050 4051 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.AclFeatureProto) 4052 } 4053 4054 static { 4055 defaultInstance = new AclFeatureProto(true); 4056 defaultInstance.initFields(); 4057 } 4058 4059 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.AclFeatureProto) 4060 } 4061 4062 public interface INodeFileOrBuilder 4063 extends com.google.protobuf.MessageOrBuilder { 4064 4065 // optional uint32 replication = 1; 4066 /** 4067 * <code>optional uint32 replication = 1;</code> 4068 */ 4069 boolean hasReplication(); 4070 /** 4071 * <code>optional uint32 replication = 1;</code> 4072 */ 4073 int getReplication(); 4074 4075 // optional uint64 modificationTime = 2; 4076 /** 4077 * <code>optional uint64 modificationTime = 2;</code> 4078 */ 4079 boolean hasModificationTime(); 4080 /** 4081 * <code>optional uint64 modificationTime = 2;</code> 4082 */ 4083 long getModificationTime(); 4084 4085 // optional uint64 accessTime = 3; 4086 /** 4087 * <code>optional uint64 accessTime = 3;</code> 4088 */ 4089 boolean hasAccessTime(); 4090 /** 4091 * <code>optional uint64 accessTime = 3;</code> 4092 */ 4093 long getAccessTime(); 4094 4095 // optional uint64 preferredBlockSize = 4; 4096 /** 4097 * <code>optional uint64 preferredBlockSize = 4;</code> 4098 */ 4099 boolean hasPreferredBlockSize(); 4100 /** 4101 * <code>optional uint64 preferredBlockSize = 4;</code> 4102 */ 4103 long getPreferredBlockSize(); 4104 4105 // optional fixed64 permission = 5; 4106 /** 4107 * <code>optional fixed64 permission = 5;</code> 4108 */ 4109 boolean hasPermission(); 4110 /** 4111 * <code>optional fixed64 permission = 5;</code> 4112 */ 4113 long getPermission(); 4114 4115 // repeated .hadoop.hdfs.BlockProto blocks = 6; 4116 /** 4117 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4118 */ 4119 java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> 4120 getBlocksList(); 4121 /** 4122 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4123 */ 4124 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index); 4125 /** 4126 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4127 */ 4128 int getBlocksCount(); 4129 /** 4130 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4131 */ 4132 java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 4133 getBlocksOrBuilderList(); 4134 /** 4135 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4136 */ 4137 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 4138 int index); 4139 4140 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 4141 /** 4142 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4143 */ 4144 boolean hasFileUC(); 4145 /** 4146 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4147 */ 4148 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC(); 4149 /** 4150 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4151 */ 4152 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder(); 4153 4154 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 4155 /** 4156 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4157 */ 4158 boolean hasAcl(); 4159 /** 4160 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4161 */ 4162 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl(); 4163 /** 4164 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4165 */ 4166 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder(); 4167 } 4168 /** 4169 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeFile} 4170 */ 4171 public static final class INodeFile extends 4172 com.google.protobuf.GeneratedMessage 4173 implements INodeFileOrBuilder { 4174 // Use INodeFile.newBuilder() to construct. 4175 private INodeFile(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4176 super(builder); 4177 this.unknownFields = builder.getUnknownFields(); 4178 } 4179 private INodeFile(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4180 4181 private static final INodeFile defaultInstance; 4182 public static INodeFile getDefaultInstance() { 4183 return defaultInstance; 4184 } 4185 4186 public INodeFile getDefaultInstanceForType() { 4187 return defaultInstance; 4188 } 4189 4190 private final com.google.protobuf.UnknownFieldSet unknownFields; 4191 @java.lang.Override 4192 public final com.google.protobuf.UnknownFieldSet 4193 getUnknownFields() { 4194 return this.unknownFields; 4195 } 4196 private INodeFile( 4197 com.google.protobuf.CodedInputStream input, 4198 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4199 throws com.google.protobuf.InvalidProtocolBufferException { 4200 initFields(); 4201 int mutable_bitField0_ = 0; 4202 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4203 com.google.protobuf.UnknownFieldSet.newBuilder(); 4204 try { 4205 boolean done = false; 4206 while (!done) { 4207 int tag = input.readTag(); 4208 switch (tag) { 4209 case 0: 4210 done = true; 4211 break; 4212 default: { 4213 if (!parseUnknownField(input, unknownFields, 4214 extensionRegistry, tag)) { 4215 done = true; 4216 } 4217 break; 4218 } 4219 case 8: { 4220 bitField0_ |= 0x00000001; 4221 replication_ = input.readUInt32(); 4222 break; 4223 } 4224 case 16: { 4225 bitField0_ |= 0x00000002; 4226 modificationTime_ = input.readUInt64(); 4227 break; 4228 } 4229 case 24: { 4230 bitField0_ |= 0x00000004; 4231 accessTime_ = input.readUInt64(); 4232 break; 4233 } 4234 case 32: { 4235 bitField0_ |= 0x00000008; 4236 preferredBlockSize_ = input.readUInt64(); 4237 break; 4238 } 4239 case 41: { 4240 bitField0_ |= 0x00000010; 4241 permission_ = input.readFixed64(); 4242 break; 4243 } 4244 case 50: { 4245 if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) { 4246 blocks_ = new java.util.ArrayList<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto>(); 4247 mutable_bitField0_ |= 0x00000020; 4248 } 4249 blocks_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.PARSER, extensionRegistry)); 4250 break; 4251 } 4252 case 58: { 4253 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder subBuilder = null; 4254 if (((bitField0_ & 0x00000020) == 0x00000020)) { 4255 subBuilder = fileUC_.toBuilder(); 4256 } 4257 fileUC_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.PARSER, extensionRegistry); 4258 if (subBuilder != null) { 4259 subBuilder.mergeFrom(fileUC_); 4260 fileUC_ = subBuilder.buildPartial(); 4261 } 4262 bitField0_ |= 0x00000020; 4263 break; 4264 } 4265 case 66: { 4266 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder subBuilder = null; 4267 if (((bitField0_ & 0x00000040) == 0x00000040)) { 4268 subBuilder = acl_.toBuilder(); 4269 } 4270 acl_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.PARSER, extensionRegistry); 4271 if (subBuilder != null) { 4272 subBuilder.mergeFrom(acl_); 4273 acl_ = subBuilder.buildPartial(); 4274 } 4275 bitField0_ |= 0x00000040; 4276 break; 4277 } 4278 } 4279 } 4280 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4281 throw e.setUnfinishedMessage(this); 4282 } catch (java.io.IOException e) { 4283 throw new com.google.protobuf.InvalidProtocolBufferException( 4284 e.getMessage()).setUnfinishedMessage(this); 4285 } finally { 4286 if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) { 4287 blocks_ = java.util.Collections.unmodifiableList(blocks_); 4288 } 4289 this.unknownFields = unknownFields.build(); 4290 makeExtensionsImmutable(); 4291 } 4292 } 4293 public static final com.google.protobuf.Descriptors.Descriptor 4294 getDescriptor() { 4295 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 4296 } 4297 4298 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4299 internalGetFieldAccessorTable() { 4300 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable 4301 .ensureFieldAccessorsInitialized( 4302 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder.class); 4303 } 4304 4305 public static com.google.protobuf.Parser<INodeFile> PARSER = 4306 new com.google.protobuf.AbstractParser<INodeFile>() { 4307 public INodeFile parsePartialFrom( 4308 com.google.protobuf.CodedInputStream input, 4309 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4310 throws com.google.protobuf.InvalidProtocolBufferException { 4311 return new INodeFile(input, extensionRegistry); 4312 } 4313 }; 4314 4315 @java.lang.Override 4316 public com.google.protobuf.Parser<INodeFile> getParserForType() { 4317 return PARSER; 4318 } 4319 4320 private int bitField0_; 4321 // optional uint32 replication = 1; 4322 public static final int REPLICATION_FIELD_NUMBER = 1; 4323 private int replication_; 4324 /** 4325 * <code>optional uint32 replication = 1;</code> 4326 */ 4327 public boolean hasReplication() { 4328 return ((bitField0_ & 0x00000001) == 0x00000001); 4329 } 4330 /** 4331 * <code>optional uint32 replication = 1;</code> 4332 */ 4333 public int getReplication() { 4334 return replication_; 4335 } 4336 4337 // optional uint64 modificationTime = 2; 4338 public static final int MODIFICATIONTIME_FIELD_NUMBER = 2; 4339 private long modificationTime_; 4340 /** 4341 * <code>optional uint64 modificationTime = 2;</code> 4342 */ 4343 public boolean hasModificationTime() { 4344 return ((bitField0_ & 0x00000002) == 0x00000002); 4345 } 4346 /** 4347 * <code>optional uint64 modificationTime = 2;</code> 4348 */ 4349 public long getModificationTime() { 4350 return modificationTime_; 4351 } 4352 4353 // optional uint64 accessTime = 3; 4354 public static final int ACCESSTIME_FIELD_NUMBER = 3; 4355 private long accessTime_; 4356 /** 4357 * <code>optional uint64 accessTime = 3;</code> 4358 */ 4359 public boolean hasAccessTime() { 4360 return ((bitField0_ & 0x00000004) == 0x00000004); 4361 } 4362 /** 4363 * <code>optional uint64 accessTime = 3;</code> 4364 */ 4365 public long getAccessTime() { 4366 return accessTime_; 4367 } 4368 4369 // optional uint64 preferredBlockSize = 4; 4370 public static final int PREFERREDBLOCKSIZE_FIELD_NUMBER = 4; 4371 private long preferredBlockSize_; 4372 /** 4373 * <code>optional uint64 preferredBlockSize = 4;</code> 4374 */ 4375 public boolean hasPreferredBlockSize() { 4376 return ((bitField0_ & 0x00000008) == 0x00000008); 4377 } 4378 /** 4379 * <code>optional uint64 preferredBlockSize = 4;</code> 4380 */ 4381 public long getPreferredBlockSize() { 4382 return preferredBlockSize_; 4383 } 4384 4385 // optional fixed64 permission = 5; 4386 public static final int PERMISSION_FIELD_NUMBER = 5; 4387 private long permission_; 4388 /** 4389 * <code>optional fixed64 permission = 5;</code> 4390 */ 4391 public boolean hasPermission() { 4392 return ((bitField0_ & 0x00000010) == 0x00000010); 4393 } 4394 /** 4395 * <code>optional fixed64 permission = 5;</code> 4396 */ 4397 public long getPermission() { 4398 return permission_; 4399 } 4400 4401 // repeated .hadoop.hdfs.BlockProto blocks = 6; 4402 public static final int BLOCKS_FIELD_NUMBER = 6; 4403 private java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> blocks_; 4404 /** 4405 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4406 */ 4407 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> getBlocksList() { 4408 return blocks_; 4409 } 4410 /** 4411 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4412 */ 4413 public java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 4414 getBlocksOrBuilderList() { 4415 return blocks_; 4416 } 4417 /** 4418 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4419 */ 4420 public int getBlocksCount() { 4421 return blocks_.size(); 4422 } 4423 /** 4424 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4425 */ 4426 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { 4427 return blocks_.get(index); 4428 } 4429 /** 4430 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 4431 */ 4432 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 4433 int index) { 4434 return blocks_.get(index); 4435 } 4436 4437 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 4438 public static final int FILEUC_FIELD_NUMBER = 7; 4439 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature fileUC_; 4440 /** 4441 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4442 */ 4443 public boolean hasFileUC() { 4444 return ((bitField0_ & 0x00000020) == 0x00000020); 4445 } 4446 /** 4447 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4448 */ 4449 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC() { 4450 return fileUC_; 4451 } 4452 /** 4453 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 4454 */ 4455 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder() { 4456 return fileUC_; 4457 } 4458 4459 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 4460 public static final int ACL_FIELD_NUMBER = 8; 4461 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_; 4462 /** 4463 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4464 */ 4465 public boolean hasAcl() { 4466 return ((bitField0_ & 0x00000040) == 0x00000040); 4467 } 4468 /** 4469 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4470 */ 4471 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 4472 return acl_; 4473 } 4474 /** 4475 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 4476 */ 4477 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 4478 return acl_; 4479 } 4480 4481 private void initFields() { 4482 replication_ = 0; 4483 modificationTime_ = 0L; 4484 accessTime_ = 0L; 4485 preferredBlockSize_ = 0L; 4486 permission_ = 0L; 4487 blocks_ = java.util.Collections.emptyList(); 4488 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 4489 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 4490 } 4491 private byte memoizedIsInitialized = -1; 4492 public final boolean isInitialized() { 4493 byte isInitialized = memoizedIsInitialized; 4494 if (isInitialized != -1) return isInitialized == 1; 4495 4496 for (int i = 0; i < getBlocksCount(); i++) { 4497 if (!getBlocks(i).isInitialized()) { 4498 memoizedIsInitialized = 0; 4499 return false; 4500 } 4501 } 4502 memoizedIsInitialized = 1; 4503 return true; 4504 } 4505 4506 public void writeTo(com.google.protobuf.CodedOutputStream output) 4507 throws java.io.IOException { 4508 getSerializedSize(); 4509 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4510 output.writeUInt32(1, replication_); 4511 } 4512 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4513 output.writeUInt64(2, modificationTime_); 4514 } 4515 if (((bitField0_ & 0x00000004) == 0x00000004)) { 4516 output.writeUInt64(3, accessTime_); 4517 } 4518 if (((bitField0_ & 0x00000008) == 0x00000008)) { 4519 output.writeUInt64(4, preferredBlockSize_); 4520 } 4521 if (((bitField0_ & 0x00000010) == 0x00000010)) { 4522 output.writeFixed64(5, permission_); 4523 } 4524 for (int i = 0; i < blocks_.size(); i++) { 4525 output.writeMessage(6, blocks_.get(i)); 4526 } 4527 if (((bitField0_ & 0x00000020) == 0x00000020)) { 4528 output.writeMessage(7, fileUC_); 4529 } 4530 if (((bitField0_ & 0x00000040) == 0x00000040)) { 4531 output.writeMessage(8, acl_); 4532 } 4533 getUnknownFields().writeTo(output); 4534 } 4535 4536 private int memoizedSerializedSize = -1; 4537 public int getSerializedSize() { 4538 int size = memoizedSerializedSize; 4539 if (size != -1) return size; 4540 4541 size = 0; 4542 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4543 size += com.google.protobuf.CodedOutputStream 4544 .computeUInt32Size(1, replication_); 4545 } 4546 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4547 size += com.google.protobuf.CodedOutputStream 4548 .computeUInt64Size(2, modificationTime_); 4549 } 4550 if (((bitField0_ & 0x00000004) == 0x00000004)) { 4551 size += com.google.protobuf.CodedOutputStream 4552 .computeUInt64Size(3, accessTime_); 4553 } 4554 if (((bitField0_ & 0x00000008) == 0x00000008)) { 4555 size += com.google.protobuf.CodedOutputStream 4556 .computeUInt64Size(4, preferredBlockSize_); 4557 } 4558 if (((bitField0_ & 0x00000010) == 0x00000010)) { 4559 size += com.google.protobuf.CodedOutputStream 4560 .computeFixed64Size(5, permission_); 4561 } 4562 for (int i = 0; i < blocks_.size(); i++) { 4563 size += com.google.protobuf.CodedOutputStream 4564 .computeMessageSize(6, blocks_.get(i)); 4565 } 4566 if (((bitField0_ & 0x00000020) == 0x00000020)) { 4567 size += com.google.protobuf.CodedOutputStream 4568 .computeMessageSize(7, fileUC_); 4569 } 4570 if (((bitField0_ & 0x00000040) == 0x00000040)) { 4571 size += com.google.protobuf.CodedOutputStream 4572 .computeMessageSize(8, acl_); 4573 } 4574 size += getUnknownFields().getSerializedSize(); 4575 memoizedSerializedSize = size; 4576 return size; 4577 } 4578 4579 private static final long serialVersionUID = 0L; 4580 @java.lang.Override 4581 protected java.lang.Object writeReplace() 4582 throws java.io.ObjectStreamException { 4583 return super.writeReplace(); 4584 } 4585 4586 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4587 com.google.protobuf.ByteString data) 4588 throws com.google.protobuf.InvalidProtocolBufferException { 4589 return PARSER.parseFrom(data); 4590 } 4591 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4592 com.google.protobuf.ByteString data, 4593 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4594 throws com.google.protobuf.InvalidProtocolBufferException { 4595 return PARSER.parseFrom(data, extensionRegistry); 4596 } 4597 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom(byte[] data) 4598 throws com.google.protobuf.InvalidProtocolBufferException { 4599 return PARSER.parseFrom(data); 4600 } 4601 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4602 byte[] data, 4603 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4604 throws com.google.protobuf.InvalidProtocolBufferException { 4605 return PARSER.parseFrom(data, extensionRegistry); 4606 } 4607 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom(java.io.InputStream input) 4608 throws java.io.IOException { 4609 return PARSER.parseFrom(input); 4610 } 4611 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4612 java.io.InputStream input, 4613 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4614 throws java.io.IOException { 4615 return PARSER.parseFrom(input, extensionRegistry); 4616 } 4617 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseDelimitedFrom(java.io.InputStream input) 4618 throws java.io.IOException { 4619 return PARSER.parseDelimitedFrom(input); 4620 } 4621 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseDelimitedFrom( 4622 java.io.InputStream input, 4623 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4624 throws java.io.IOException { 4625 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4626 } 4627 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4628 com.google.protobuf.CodedInputStream input) 4629 throws java.io.IOException { 4630 return PARSER.parseFrom(input); 4631 } 4632 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 4633 com.google.protobuf.CodedInputStream input, 4634 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4635 throws java.io.IOException { 4636 return PARSER.parseFrom(input, extensionRegistry); 4637 } 4638 4639 public static Builder newBuilder() { return Builder.create(); } 4640 public Builder newBuilderForType() { return newBuilder(); } 4641 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile prototype) { 4642 return newBuilder().mergeFrom(prototype); 4643 } 4644 public Builder toBuilder() { return newBuilder(this); } 4645 4646 @java.lang.Override 4647 protected Builder newBuilderForType( 4648 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4649 Builder builder = new Builder(parent); 4650 return builder; 4651 } 4652 /** 4653 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeFile} 4654 */ 4655 public static final class Builder extends 4656 com.google.protobuf.GeneratedMessage.Builder<Builder> 4657 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder { 4658 public static final com.google.protobuf.Descriptors.Descriptor 4659 getDescriptor() { 4660 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 4661 } 4662 4663 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4664 internalGetFieldAccessorTable() { 4665 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable 4666 .ensureFieldAccessorsInitialized( 4667 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder.class); 4668 } 4669 4670 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder() 4671 private Builder() { 4672 maybeForceBuilderInitialization(); 4673 } 4674 4675 private Builder( 4676 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4677 super(parent); 4678 maybeForceBuilderInitialization(); 4679 } 4680 private void maybeForceBuilderInitialization() { 4681 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4682 getBlocksFieldBuilder(); 4683 getFileUCFieldBuilder(); 4684 getAclFieldBuilder(); 4685 } 4686 } 4687 private static Builder create() { 4688 return new Builder(); 4689 } 4690 4691 public Builder clear() { 4692 super.clear(); 4693 replication_ = 0; 4694 bitField0_ = (bitField0_ & ~0x00000001); 4695 modificationTime_ = 0L; 4696 bitField0_ = (bitField0_ & ~0x00000002); 4697 accessTime_ = 0L; 4698 bitField0_ = (bitField0_ & ~0x00000004); 4699 preferredBlockSize_ = 0L; 4700 bitField0_ = (bitField0_ & ~0x00000008); 4701 permission_ = 0L; 4702 bitField0_ = (bitField0_ & ~0x00000010); 4703 if (blocksBuilder_ == null) { 4704 blocks_ = java.util.Collections.emptyList(); 4705 bitField0_ = (bitField0_ & ~0x00000020); 4706 } else { 4707 blocksBuilder_.clear(); 4708 } 4709 if (fileUCBuilder_ == null) { 4710 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 4711 } else { 4712 fileUCBuilder_.clear(); 4713 } 4714 bitField0_ = (bitField0_ & ~0x00000040); 4715 if (aclBuilder_ == null) { 4716 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 4717 } else { 4718 aclBuilder_.clear(); 4719 } 4720 bitField0_ = (bitField0_ & ~0x00000080); 4721 return this; 4722 } 4723 4724 public Builder clone() { 4725 return create().mergeFrom(buildPartial()); 4726 } 4727 4728 public com.google.protobuf.Descriptors.Descriptor 4729 getDescriptorForType() { 4730 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 4731 } 4732 4733 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getDefaultInstanceForType() { 4734 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 4735 } 4736 4737 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile build() { 4738 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile result = buildPartial(); 4739 if (!result.isInitialized()) { 4740 throw newUninitializedMessageException(result); 4741 } 4742 return result; 4743 } 4744 4745 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile buildPartial() { 4746 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile(this); 4747 int from_bitField0_ = bitField0_; 4748 int to_bitField0_ = 0; 4749 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 4750 to_bitField0_ |= 0x00000001; 4751 } 4752 result.replication_ = replication_; 4753 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 4754 to_bitField0_ |= 0x00000002; 4755 } 4756 result.modificationTime_ = modificationTime_; 4757 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 4758 to_bitField0_ |= 0x00000004; 4759 } 4760 result.accessTime_ = accessTime_; 4761 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 4762 to_bitField0_ |= 0x00000008; 4763 } 4764 result.preferredBlockSize_ = preferredBlockSize_; 4765 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 4766 to_bitField0_ |= 0x00000010; 4767 } 4768 result.permission_ = permission_; 4769 if (blocksBuilder_ == null) { 4770 if (((bitField0_ & 0x00000020) == 0x00000020)) { 4771 blocks_ = java.util.Collections.unmodifiableList(blocks_); 4772 bitField0_ = (bitField0_ & ~0x00000020); 4773 } 4774 result.blocks_ = blocks_; 4775 } else { 4776 result.blocks_ = blocksBuilder_.build(); 4777 } 4778 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 4779 to_bitField0_ |= 0x00000020; 4780 } 4781 if (fileUCBuilder_ == null) { 4782 result.fileUC_ = fileUC_; 4783 } else { 4784 result.fileUC_ = fileUCBuilder_.build(); 4785 } 4786 if (((from_bitField0_ & 0x00000080) == 0x00000080)) { 4787 to_bitField0_ |= 0x00000040; 4788 } 4789 if (aclBuilder_ == null) { 4790 result.acl_ = acl_; 4791 } else { 4792 result.acl_ = aclBuilder_.build(); 4793 } 4794 result.bitField0_ = to_bitField0_; 4795 onBuilt(); 4796 return result; 4797 } 4798 4799 public Builder mergeFrom(com.google.protobuf.Message other) { 4800 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile) { 4801 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile)other); 4802 } else { 4803 super.mergeFrom(other); 4804 return this; 4805 } 4806 } 4807 4808 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile other) { 4809 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) return this; 4810 if (other.hasReplication()) { 4811 setReplication(other.getReplication()); 4812 } 4813 if (other.hasModificationTime()) { 4814 setModificationTime(other.getModificationTime()); 4815 } 4816 if (other.hasAccessTime()) { 4817 setAccessTime(other.getAccessTime()); 4818 } 4819 if (other.hasPreferredBlockSize()) { 4820 setPreferredBlockSize(other.getPreferredBlockSize()); 4821 } 4822 if (other.hasPermission()) { 4823 setPermission(other.getPermission()); 4824 } 4825 if (blocksBuilder_ == null) { 4826 if (!other.blocks_.isEmpty()) { 4827 if (blocks_.isEmpty()) { 4828 blocks_ = other.blocks_; 4829 bitField0_ = (bitField0_ & ~0x00000020); 4830 } else { 4831 ensureBlocksIsMutable(); 4832 blocks_.addAll(other.blocks_); 4833 } 4834 onChanged(); 4835 } 4836 } else { 4837 if (!other.blocks_.isEmpty()) { 4838 if (blocksBuilder_.isEmpty()) { 4839 blocksBuilder_.dispose(); 4840 blocksBuilder_ = null; 4841 blocks_ = other.blocks_; 4842 bitField0_ = (bitField0_ & ~0x00000020); 4843 blocksBuilder_ = 4844 com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? 4845 getBlocksFieldBuilder() : null; 4846 } else { 4847 blocksBuilder_.addAllMessages(other.blocks_); 4848 } 4849 } 4850 } 4851 if (other.hasFileUC()) { 4852 mergeFileUC(other.getFileUC()); 4853 } 4854 if (other.hasAcl()) { 4855 mergeAcl(other.getAcl()); 4856 } 4857 this.mergeUnknownFields(other.getUnknownFields()); 4858 return this; 4859 } 4860 4861 public final boolean isInitialized() { 4862 for (int i = 0; i < getBlocksCount(); i++) { 4863 if (!getBlocks(i).isInitialized()) { 4864 4865 return false; 4866 } 4867 } 4868 return true; 4869 } 4870 4871 public Builder mergeFrom( 4872 com.google.protobuf.CodedInputStream input, 4873 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4874 throws java.io.IOException { 4875 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parsedMessage = null; 4876 try { 4877 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4878 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4879 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile) e.getUnfinishedMessage(); 4880 throw e; 4881 } finally { 4882 if (parsedMessage != null) { 4883 mergeFrom(parsedMessage); 4884 } 4885 } 4886 return this; 4887 } 4888 private int bitField0_; 4889 4890 // optional uint32 replication = 1; 4891 private int replication_ ; 4892 /** 4893 * <code>optional uint32 replication = 1;</code> 4894 */ 4895 public boolean hasReplication() { 4896 return ((bitField0_ & 0x00000001) == 0x00000001); 4897 } 4898 /** 4899 * <code>optional uint32 replication = 1;</code> 4900 */ 4901 public int getReplication() { 4902 return replication_; 4903 } 4904 /** 4905 * <code>optional uint32 replication = 1;</code> 4906 */ 4907 public Builder setReplication(int value) { 4908 bitField0_ |= 0x00000001; 4909 replication_ = value; 4910 onChanged(); 4911 return this; 4912 } 4913 /** 4914 * <code>optional uint32 replication = 1;</code> 4915 */ 4916 public Builder clearReplication() { 4917 bitField0_ = (bitField0_ & ~0x00000001); 4918 replication_ = 0; 4919 onChanged(); 4920 return this; 4921 } 4922 4923 // optional uint64 modificationTime = 2; 4924 private long modificationTime_ ; 4925 /** 4926 * <code>optional uint64 modificationTime = 2;</code> 4927 */ 4928 public boolean hasModificationTime() { 4929 return ((bitField0_ & 0x00000002) == 0x00000002); 4930 } 4931 /** 4932 * <code>optional uint64 modificationTime = 2;</code> 4933 */ 4934 public long getModificationTime() { 4935 return modificationTime_; 4936 } 4937 /** 4938 * <code>optional uint64 modificationTime = 2;</code> 4939 */ 4940 public Builder setModificationTime(long value) { 4941 bitField0_ |= 0x00000002; 4942 modificationTime_ = value; 4943 onChanged(); 4944 return this; 4945 } 4946 /** 4947 * <code>optional uint64 modificationTime = 2;</code> 4948 */ 4949 public Builder clearModificationTime() { 4950 bitField0_ = (bitField0_ & ~0x00000002); 4951 modificationTime_ = 0L; 4952 onChanged(); 4953 return this; 4954 } 4955 4956 // optional uint64 accessTime = 3; 4957 private long accessTime_ ; 4958 /** 4959 * <code>optional uint64 accessTime = 3;</code> 4960 */ 4961 public boolean hasAccessTime() { 4962 return ((bitField0_ & 0x00000004) == 0x00000004); 4963 } 4964 /** 4965 * <code>optional uint64 accessTime = 3;</code> 4966 */ 4967 public long getAccessTime() { 4968 return accessTime_; 4969 } 4970 /** 4971 * <code>optional uint64 accessTime = 3;</code> 4972 */ 4973 public Builder setAccessTime(long value) { 4974 bitField0_ |= 0x00000004; 4975 accessTime_ = value; 4976 onChanged(); 4977 return this; 4978 } 4979 /** 4980 * <code>optional uint64 accessTime = 3;</code> 4981 */ 4982 public Builder clearAccessTime() { 4983 bitField0_ = (bitField0_ & ~0x00000004); 4984 accessTime_ = 0L; 4985 onChanged(); 4986 return this; 4987 } 4988 4989 // optional uint64 preferredBlockSize = 4; 4990 private long preferredBlockSize_ ; 4991 /** 4992 * <code>optional uint64 preferredBlockSize = 4;</code> 4993 */ 4994 public boolean hasPreferredBlockSize() { 4995 return ((bitField0_ & 0x00000008) == 0x00000008); 4996 } 4997 /** 4998 * <code>optional uint64 preferredBlockSize = 4;</code> 4999 */ 5000 public long getPreferredBlockSize() { 5001 return preferredBlockSize_; 5002 } 5003 /** 5004 * <code>optional uint64 preferredBlockSize = 4;</code> 5005 */ 5006 public Builder setPreferredBlockSize(long value) { 5007 bitField0_ |= 0x00000008; 5008 preferredBlockSize_ = value; 5009 onChanged(); 5010 return this; 5011 } 5012 /** 5013 * <code>optional uint64 preferredBlockSize = 4;</code> 5014 */ 5015 public Builder clearPreferredBlockSize() { 5016 bitField0_ = (bitField0_ & ~0x00000008); 5017 preferredBlockSize_ = 0L; 5018 onChanged(); 5019 return this; 5020 } 5021 5022 // optional fixed64 permission = 5; 5023 private long permission_ ; 5024 /** 5025 * <code>optional fixed64 permission = 5;</code> 5026 */ 5027 public boolean hasPermission() { 5028 return ((bitField0_ & 0x00000010) == 0x00000010); 5029 } 5030 /** 5031 * <code>optional fixed64 permission = 5;</code> 5032 */ 5033 public long getPermission() { 5034 return permission_; 5035 } 5036 /** 5037 * <code>optional fixed64 permission = 5;</code> 5038 */ 5039 public Builder setPermission(long value) { 5040 bitField0_ |= 0x00000010; 5041 permission_ = value; 5042 onChanged(); 5043 return this; 5044 } 5045 /** 5046 * <code>optional fixed64 permission = 5;</code> 5047 */ 5048 public Builder clearPermission() { 5049 bitField0_ = (bitField0_ & ~0x00000010); 5050 permission_ = 0L; 5051 onChanged(); 5052 return this; 5053 } 5054 5055 // repeated .hadoop.hdfs.BlockProto blocks = 6; 5056 private java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> blocks_ = 5057 java.util.Collections.emptyList(); 5058 private void ensureBlocksIsMutable() { 5059 if (!((bitField0_ & 0x00000020) == 0x00000020)) { 5060 blocks_ = new java.util.ArrayList<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto>(blocks_); 5061 bitField0_ |= 0x00000020; 5062 } 5063 } 5064 5065 private com.google.protobuf.RepeatedFieldBuilder< 5066 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> blocksBuilder_; 5067 5068 /** 5069 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5070 */ 5071 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> getBlocksList() { 5072 if (blocksBuilder_ == null) { 5073 return java.util.Collections.unmodifiableList(blocks_); 5074 } else { 5075 return blocksBuilder_.getMessageList(); 5076 } 5077 } 5078 /** 5079 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5080 */ 5081 public int getBlocksCount() { 5082 if (blocksBuilder_ == null) { 5083 return blocks_.size(); 5084 } else { 5085 return blocksBuilder_.getCount(); 5086 } 5087 } 5088 /** 5089 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5090 */ 5091 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { 5092 if (blocksBuilder_ == null) { 5093 return blocks_.get(index); 5094 } else { 5095 return blocksBuilder_.getMessage(index); 5096 } 5097 } 5098 /** 5099 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5100 */ 5101 public Builder setBlocks( 5102 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 5103 if (blocksBuilder_ == null) { 5104 if (value == null) { 5105 throw new NullPointerException(); 5106 } 5107 ensureBlocksIsMutable(); 5108 blocks_.set(index, value); 5109 onChanged(); 5110 } else { 5111 blocksBuilder_.setMessage(index, value); 5112 } 5113 return this; 5114 } 5115 /** 5116 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5117 */ 5118 public Builder setBlocks( 5119 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 5120 if (blocksBuilder_ == null) { 5121 ensureBlocksIsMutable(); 5122 blocks_.set(index, builderForValue.build()); 5123 onChanged(); 5124 } else { 5125 blocksBuilder_.setMessage(index, builderForValue.build()); 5126 } 5127 return this; 5128 } 5129 /** 5130 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5131 */ 5132 public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 5133 if (blocksBuilder_ == null) { 5134 if (value == null) { 5135 throw new NullPointerException(); 5136 } 5137 ensureBlocksIsMutable(); 5138 blocks_.add(value); 5139 onChanged(); 5140 } else { 5141 blocksBuilder_.addMessage(value); 5142 } 5143 return this; 5144 } 5145 /** 5146 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5147 */ 5148 public Builder addBlocks( 5149 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 5150 if (blocksBuilder_ == null) { 5151 if (value == null) { 5152 throw new NullPointerException(); 5153 } 5154 ensureBlocksIsMutable(); 5155 blocks_.add(index, value); 5156 onChanged(); 5157 } else { 5158 blocksBuilder_.addMessage(index, value); 5159 } 5160 return this; 5161 } 5162 /** 5163 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5164 */ 5165 public Builder addBlocks( 5166 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 5167 if (blocksBuilder_ == null) { 5168 ensureBlocksIsMutable(); 5169 blocks_.add(builderForValue.build()); 5170 onChanged(); 5171 } else { 5172 blocksBuilder_.addMessage(builderForValue.build()); 5173 } 5174 return this; 5175 } 5176 /** 5177 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5178 */ 5179 public Builder addBlocks( 5180 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 5181 if (blocksBuilder_ == null) { 5182 ensureBlocksIsMutable(); 5183 blocks_.add(index, builderForValue.build()); 5184 onChanged(); 5185 } else { 5186 blocksBuilder_.addMessage(index, builderForValue.build()); 5187 } 5188 return this; 5189 } 5190 /** 5191 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5192 */ 5193 public Builder addAllBlocks( 5194 java.lang.Iterable<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> values) { 5195 if (blocksBuilder_ == null) { 5196 ensureBlocksIsMutable(); 5197 super.addAll(values, blocks_); 5198 onChanged(); 5199 } else { 5200 blocksBuilder_.addAllMessages(values); 5201 } 5202 return this; 5203 } 5204 /** 5205 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5206 */ 5207 public Builder clearBlocks() { 5208 if (blocksBuilder_ == null) { 5209 blocks_ = java.util.Collections.emptyList(); 5210 bitField0_ = (bitField0_ & ~0x00000020); 5211 onChanged(); 5212 } else { 5213 blocksBuilder_.clear(); 5214 } 5215 return this; 5216 } 5217 /** 5218 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5219 */ 5220 public Builder removeBlocks(int index) { 5221 if (blocksBuilder_ == null) { 5222 ensureBlocksIsMutable(); 5223 blocks_.remove(index); 5224 onChanged(); 5225 } else { 5226 blocksBuilder_.remove(index); 5227 } 5228 return this; 5229 } 5230 /** 5231 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5232 */ 5233 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder getBlocksBuilder( 5234 int index) { 5235 return getBlocksFieldBuilder().getBuilder(index); 5236 } 5237 /** 5238 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5239 */ 5240 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 5241 int index) { 5242 if (blocksBuilder_ == null) { 5243 return blocks_.get(index); } else { 5244 return blocksBuilder_.getMessageOrBuilder(index); 5245 } 5246 } 5247 /** 5248 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5249 */ 5250 public java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 5251 getBlocksOrBuilderList() { 5252 if (blocksBuilder_ != null) { 5253 return blocksBuilder_.getMessageOrBuilderList(); 5254 } else { 5255 return java.util.Collections.unmodifiableList(blocks_); 5256 } 5257 } 5258 /** 5259 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5260 */ 5261 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { 5262 return getBlocksFieldBuilder().addBuilder( 5263 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); 5264 } 5265 /** 5266 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5267 */ 5268 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( 5269 int index) { 5270 return getBlocksFieldBuilder().addBuilder( 5271 index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); 5272 } 5273 /** 5274 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5275 */ 5276 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder> 5277 getBlocksBuilderList() { 5278 return getBlocksFieldBuilder().getBuilderList(); 5279 } 5280 private com.google.protobuf.RepeatedFieldBuilder< 5281 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 5282 getBlocksFieldBuilder() { 5283 if (blocksBuilder_ == null) { 5284 blocksBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< 5285 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder>( 5286 blocks_, 5287 ((bitField0_ & 0x00000020) == 0x00000020), 5288 getParentForChildren(), 5289 isClean()); 5290 blocks_ = null; 5291 } 5292 return blocksBuilder_; 5293 } 5294 5295 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 5296 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 5297 private com.google.protobuf.SingleFieldBuilder< 5298 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder> fileUCBuilder_; 5299 /** 5300 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5301 */ 5302 public boolean hasFileUC() { 5303 return ((bitField0_ & 0x00000040) == 0x00000040); 5304 } 5305 /** 5306 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5307 */ 5308 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC() { 5309 if (fileUCBuilder_ == null) { 5310 return fileUC_; 5311 } else { 5312 return fileUCBuilder_.getMessage(); 5313 } 5314 } 5315 /** 5316 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5317 */ 5318 public Builder setFileUC(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature value) { 5319 if (fileUCBuilder_ == null) { 5320 if (value == null) { 5321 throw new NullPointerException(); 5322 } 5323 fileUC_ = value; 5324 onChanged(); 5325 } else { 5326 fileUCBuilder_.setMessage(value); 5327 } 5328 bitField0_ |= 0x00000040; 5329 return this; 5330 } 5331 /** 5332 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5333 */ 5334 public Builder setFileUC( 5335 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder builderForValue) { 5336 if (fileUCBuilder_ == null) { 5337 fileUC_ = builderForValue.build(); 5338 onChanged(); 5339 } else { 5340 fileUCBuilder_.setMessage(builderForValue.build()); 5341 } 5342 bitField0_ |= 0x00000040; 5343 return this; 5344 } 5345 /** 5346 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5347 */ 5348 public Builder mergeFileUC(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature value) { 5349 if (fileUCBuilder_ == null) { 5350 if (((bitField0_ & 0x00000040) == 0x00000040) && 5351 fileUC_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance()) { 5352 fileUC_ = 5353 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.newBuilder(fileUC_).mergeFrom(value).buildPartial(); 5354 } else { 5355 fileUC_ = value; 5356 } 5357 onChanged(); 5358 } else { 5359 fileUCBuilder_.mergeFrom(value); 5360 } 5361 bitField0_ |= 0x00000040; 5362 return this; 5363 } 5364 /** 5365 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5366 */ 5367 public Builder clearFileUC() { 5368 if (fileUCBuilder_ == null) { 5369 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 5370 onChanged(); 5371 } else { 5372 fileUCBuilder_.clear(); 5373 } 5374 bitField0_ = (bitField0_ & ~0x00000040); 5375 return this; 5376 } 5377 /** 5378 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5379 */ 5380 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder getFileUCBuilder() { 5381 bitField0_ |= 0x00000040; 5382 onChanged(); 5383 return getFileUCFieldBuilder().getBuilder(); 5384 } 5385 /** 5386 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5387 */ 5388 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder() { 5389 if (fileUCBuilder_ != null) { 5390 return fileUCBuilder_.getMessageOrBuilder(); 5391 } else { 5392 return fileUC_; 5393 } 5394 } 5395 /** 5396 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5397 */ 5398 private com.google.protobuf.SingleFieldBuilder< 5399 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder> 5400 getFileUCFieldBuilder() { 5401 if (fileUCBuilder_ == null) { 5402 fileUCBuilder_ = new com.google.protobuf.SingleFieldBuilder< 5403 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder>( 5404 fileUC_, 5405 getParentForChildren(), 5406 isClean()); 5407 fileUC_ = null; 5408 } 5409 return fileUCBuilder_; 5410 } 5411 5412 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 5413 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 5414 private com.google.protobuf.SingleFieldBuilder< 5415 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> aclBuilder_; 5416 /** 5417 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5418 */ 5419 public boolean hasAcl() { 5420 return ((bitField0_ & 0x00000080) == 0x00000080); 5421 } 5422 /** 5423 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5424 */ 5425 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 5426 if (aclBuilder_ == null) { 5427 return acl_; 5428 } else { 5429 return aclBuilder_.getMessage(); 5430 } 5431 } 5432 /** 5433 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5434 */ 5435 public Builder setAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 5436 if (aclBuilder_ == null) { 5437 if (value == null) { 5438 throw new NullPointerException(); 5439 } 5440 acl_ = value; 5441 onChanged(); 5442 } else { 5443 aclBuilder_.setMessage(value); 5444 } 5445 bitField0_ |= 0x00000080; 5446 return this; 5447 } 5448 /** 5449 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5450 */ 5451 public Builder setAcl( 5452 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder builderForValue) { 5453 if (aclBuilder_ == null) { 5454 acl_ = builderForValue.build(); 5455 onChanged(); 5456 } else { 5457 aclBuilder_.setMessage(builderForValue.build()); 5458 } 5459 bitField0_ |= 0x00000080; 5460 return this; 5461 } 5462 /** 5463 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5464 */ 5465 public Builder mergeAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 5466 if (aclBuilder_ == null) { 5467 if (((bitField0_ & 0x00000080) == 0x00000080) && 5468 acl_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) { 5469 acl_ = 5470 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder(acl_).mergeFrom(value).buildPartial(); 5471 } else { 5472 acl_ = value; 5473 } 5474 onChanged(); 5475 } else { 5476 aclBuilder_.mergeFrom(value); 5477 } 5478 bitField0_ |= 0x00000080; 5479 return this; 5480 } 5481 /** 5482 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5483 */ 5484 public Builder clearAcl() { 5485 if (aclBuilder_ == null) { 5486 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 5487 onChanged(); 5488 } else { 5489 aclBuilder_.clear(); 5490 } 5491 bitField0_ = (bitField0_ & ~0x00000080); 5492 return this; 5493 } 5494 /** 5495 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5496 */ 5497 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder getAclBuilder() { 5498 bitField0_ |= 0x00000080; 5499 onChanged(); 5500 return getAclFieldBuilder().getBuilder(); 5501 } 5502 /** 5503 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5504 */ 5505 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 5506 if (aclBuilder_ != null) { 5507 return aclBuilder_.getMessageOrBuilder(); 5508 } else { 5509 return acl_; 5510 } 5511 } 5512 /** 5513 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5514 */ 5515 private com.google.protobuf.SingleFieldBuilder< 5516 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> 5517 getAclFieldBuilder() { 5518 if (aclBuilder_ == null) { 5519 aclBuilder_ = new com.google.protobuf.SingleFieldBuilder< 5520 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder>( 5521 acl_, 5522 getParentForChildren(), 5523 isClean()); 5524 acl_ = null; 5525 } 5526 return aclBuilder_; 5527 } 5528 5529 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeFile) 5530 } 5531 5532 static { 5533 defaultInstance = new INodeFile(true); 5534 defaultInstance.initFields(); 5535 } 5536 5537 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeFile) 5538 } 5539 5540 public interface INodeDirectoryOrBuilder 5541 extends com.google.protobuf.MessageOrBuilder { 5542 5543 // optional uint64 modificationTime = 1; 5544 /** 5545 * <code>optional uint64 modificationTime = 1;</code> 5546 */ 5547 boolean hasModificationTime(); 5548 /** 5549 * <code>optional uint64 modificationTime = 1;</code> 5550 */ 5551 long getModificationTime(); 5552 5553 // optional uint64 nsQuota = 2; 5554 /** 5555 * <code>optional uint64 nsQuota = 2;</code> 5556 * 5557 * <pre> 5558 * namespace quota 5559 * </pre> 5560 */ 5561 boolean hasNsQuota(); 5562 /** 5563 * <code>optional uint64 nsQuota = 2;</code> 5564 * 5565 * <pre> 5566 * namespace quota 5567 * </pre> 5568 */ 5569 long getNsQuota(); 5570 5571 // optional uint64 dsQuota = 3; 5572 /** 5573 * <code>optional uint64 dsQuota = 3;</code> 5574 * 5575 * <pre> 5576 * diskspace quota 5577 * </pre> 5578 */ 5579 boolean hasDsQuota(); 5580 /** 5581 * <code>optional uint64 dsQuota = 3;</code> 5582 * 5583 * <pre> 5584 * diskspace quota 5585 * </pre> 5586 */ 5587 long getDsQuota(); 5588 5589 // optional fixed64 permission = 4; 5590 /** 5591 * <code>optional fixed64 permission = 4;</code> 5592 */ 5593 boolean hasPermission(); 5594 /** 5595 * <code>optional fixed64 permission = 4;</code> 5596 */ 5597 long getPermission(); 5598 5599 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 5600 /** 5601 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5602 */ 5603 boolean hasAcl(); 5604 /** 5605 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5606 */ 5607 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl(); 5608 /** 5609 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5610 */ 5611 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder(); 5612 } 5613 /** 5614 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeDirectory} 5615 */ 5616 public static final class INodeDirectory extends 5617 com.google.protobuf.GeneratedMessage 5618 implements INodeDirectoryOrBuilder { 5619 // Use INodeDirectory.newBuilder() to construct. 5620 private INodeDirectory(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 5621 super(builder); 5622 this.unknownFields = builder.getUnknownFields(); 5623 } 5624 private INodeDirectory(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 5625 5626 private static final INodeDirectory defaultInstance; 5627 public static INodeDirectory getDefaultInstance() { 5628 return defaultInstance; 5629 } 5630 5631 public INodeDirectory getDefaultInstanceForType() { 5632 return defaultInstance; 5633 } 5634 5635 private final com.google.protobuf.UnknownFieldSet unknownFields; 5636 @java.lang.Override 5637 public final com.google.protobuf.UnknownFieldSet 5638 getUnknownFields() { 5639 return this.unknownFields; 5640 } 5641 private INodeDirectory( 5642 com.google.protobuf.CodedInputStream input, 5643 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5644 throws com.google.protobuf.InvalidProtocolBufferException { 5645 initFields(); 5646 int mutable_bitField0_ = 0; 5647 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 5648 com.google.protobuf.UnknownFieldSet.newBuilder(); 5649 try { 5650 boolean done = false; 5651 while (!done) { 5652 int tag = input.readTag(); 5653 switch (tag) { 5654 case 0: 5655 done = true; 5656 break; 5657 default: { 5658 if (!parseUnknownField(input, unknownFields, 5659 extensionRegistry, tag)) { 5660 done = true; 5661 } 5662 break; 5663 } 5664 case 8: { 5665 bitField0_ |= 0x00000001; 5666 modificationTime_ = input.readUInt64(); 5667 break; 5668 } 5669 case 16: { 5670 bitField0_ |= 0x00000002; 5671 nsQuota_ = input.readUInt64(); 5672 break; 5673 } 5674 case 24: { 5675 bitField0_ |= 0x00000004; 5676 dsQuota_ = input.readUInt64(); 5677 break; 5678 } 5679 case 33: { 5680 bitField0_ |= 0x00000008; 5681 permission_ = input.readFixed64(); 5682 break; 5683 } 5684 case 42: { 5685 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder subBuilder = null; 5686 if (((bitField0_ & 0x00000010) == 0x00000010)) { 5687 subBuilder = acl_.toBuilder(); 5688 } 5689 acl_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.PARSER, extensionRegistry); 5690 if (subBuilder != null) { 5691 subBuilder.mergeFrom(acl_); 5692 acl_ = subBuilder.buildPartial(); 5693 } 5694 bitField0_ |= 0x00000010; 5695 break; 5696 } 5697 } 5698 } 5699 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5700 throw e.setUnfinishedMessage(this); 5701 } catch (java.io.IOException e) { 5702 throw new com.google.protobuf.InvalidProtocolBufferException( 5703 e.getMessage()).setUnfinishedMessage(this); 5704 } finally { 5705 this.unknownFields = unknownFields.build(); 5706 makeExtensionsImmutable(); 5707 } 5708 } 5709 public static final com.google.protobuf.Descriptors.Descriptor 5710 getDescriptor() { 5711 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 5712 } 5713 5714 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5715 internalGetFieldAccessorTable() { 5716 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable 5717 .ensureFieldAccessorsInitialized( 5718 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder.class); 5719 } 5720 5721 public static com.google.protobuf.Parser<INodeDirectory> PARSER = 5722 new com.google.protobuf.AbstractParser<INodeDirectory>() { 5723 public INodeDirectory parsePartialFrom( 5724 com.google.protobuf.CodedInputStream input, 5725 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5726 throws com.google.protobuf.InvalidProtocolBufferException { 5727 return new INodeDirectory(input, extensionRegistry); 5728 } 5729 }; 5730 5731 @java.lang.Override 5732 public com.google.protobuf.Parser<INodeDirectory> getParserForType() { 5733 return PARSER; 5734 } 5735 5736 private int bitField0_; 5737 // optional uint64 modificationTime = 1; 5738 public static final int MODIFICATIONTIME_FIELD_NUMBER = 1; 5739 private long modificationTime_; 5740 /** 5741 * <code>optional uint64 modificationTime = 1;</code> 5742 */ 5743 public boolean hasModificationTime() { 5744 return ((bitField0_ & 0x00000001) == 0x00000001); 5745 } 5746 /** 5747 * <code>optional uint64 modificationTime = 1;</code> 5748 */ 5749 public long getModificationTime() { 5750 return modificationTime_; 5751 } 5752 5753 // optional uint64 nsQuota = 2; 5754 public static final int NSQUOTA_FIELD_NUMBER = 2; 5755 private long nsQuota_; 5756 /** 5757 * <code>optional uint64 nsQuota = 2;</code> 5758 * 5759 * <pre> 5760 * namespace quota 5761 * </pre> 5762 */ 5763 public boolean hasNsQuota() { 5764 return ((bitField0_ & 0x00000002) == 0x00000002); 5765 } 5766 /** 5767 * <code>optional uint64 nsQuota = 2;</code> 5768 * 5769 * <pre> 5770 * namespace quota 5771 * </pre> 5772 */ 5773 public long getNsQuota() { 5774 return nsQuota_; 5775 } 5776 5777 // optional uint64 dsQuota = 3; 5778 public static final int DSQUOTA_FIELD_NUMBER = 3; 5779 private long dsQuota_; 5780 /** 5781 * <code>optional uint64 dsQuota = 3;</code> 5782 * 5783 * <pre> 5784 * diskspace quota 5785 * </pre> 5786 */ 5787 public boolean hasDsQuota() { 5788 return ((bitField0_ & 0x00000004) == 0x00000004); 5789 } 5790 /** 5791 * <code>optional uint64 dsQuota = 3;</code> 5792 * 5793 * <pre> 5794 * diskspace quota 5795 * </pre> 5796 */ 5797 public long getDsQuota() { 5798 return dsQuota_; 5799 } 5800 5801 // optional fixed64 permission = 4; 5802 public static final int PERMISSION_FIELD_NUMBER = 4; 5803 private long permission_; 5804 /** 5805 * <code>optional fixed64 permission = 4;</code> 5806 */ 5807 public boolean hasPermission() { 5808 return ((bitField0_ & 0x00000008) == 0x00000008); 5809 } 5810 /** 5811 * <code>optional fixed64 permission = 4;</code> 5812 */ 5813 public long getPermission() { 5814 return permission_; 5815 } 5816 5817 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 5818 public static final int ACL_FIELD_NUMBER = 5; 5819 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_; 5820 /** 5821 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5822 */ 5823 public boolean hasAcl() { 5824 return ((bitField0_ & 0x00000010) == 0x00000010); 5825 } 5826 /** 5827 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5828 */ 5829 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 5830 return acl_; 5831 } 5832 /** 5833 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 5834 */ 5835 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 5836 return acl_; 5837 } 5838 5839 private void initFields() { 5840 modificationTime_ = 0L; 5841 nsQuota_ = 0L; 5842 dsQuota_ = 0L; 5843 permission_ = 0L; 5844 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 5845 } 5846 private byte memoizedIsInitialized = -1; 5847 public final boolean isInitialized() { 5848 byte isInitialized = memoizedIsInitialized; 5849 if (isInitialized != -1) return isInitialized == 1; 5850 5851 memoizedIsInitialized = 1; 5852 return true; 5853 } 5854 5855 public void writeTo(com.google.protobuf.CodedOutputStream output) 5856 throws java.io.IOException { 5857 getSerializedSize(); 5858 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5859 output.writeUInt64(1, modificationTime_); 5860 } 5861 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5862 output.writeUInt64(2, nsQuota_); 5863 } 5864 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5865 output.writeUInt64(3, dsQuota_); 5866 } 5867 if (((bitField0_ & 0x00000008) == 0x00000008)) { 5868 output.writeFixed64(4, permission_); 5869 } 5870 if (((bitField0_ & 0x00000010) == 0x00000010)) { 5871 output.writeMessage(5, acl_); 5872 } 5873 getUnknownFields().writeTo(output); 5874 } 5875 5876 private int memoizedSerializedSize = -1; 5877 public int getSerializedSize() { 5878 int size = memoizedSerializedSize; 5879 if (size != -1) return size; 5880 5881 size = 0; 5882 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5883 size += com.google.protobuf.CodedOutputStream 5884 .computeUInt64Size(1, modificationTime_); 5885 } 5886 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5887 size += com.google.protobuf.CodedOutputStream 5888 .computeUInt64Size(2, nsQuota_); 5889 } 5890 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5891 size += com.google.protobuf.CodedOutputStream 5892 .computeUInt64Size(3, dsQuota_); 5893 } 5894 if (((bitField0_ & 0x00000008) == 0x00000008)) { 5895 size += com.google.protobuf.CodedOutputStream 5896 .computeFixed64Size(4, permission_); 5897 } 5898 if (((bitField0_ & 0x00000010) == 0x00000010)) { 5899 size += com.google.protobuf.CodedOutputStream 5900 .computeMessageSize(5, acl_); 5901 } 5902 size += getUnknownFields().getSerializedSize(); 5903 memoizedSerializedSize = size; 5904 return size; 5905 } 5906 5907 private static final long serialVersionUID = 0L; 5908 @java.lang.Override 5909 protected java.lang.Object writeReplace() 5910 throws java.io.ObjectStreamException { 5911 return super.writeReplace(); 5912 } 5913 5914 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5915 com.google.protobuf.ByteString data) 5916 throws com.google.protobuf.InvalidProtocolBufferException { 5917 return PARSER.parseFrom(data); 5918 } 5919 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5920 com.google.protobuf.ByteString data, 5921 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5922 throws com.google.protobuf.InvalidProtocolBufferException { 5923 return PARSER.parseFrom(data, extensionRegistry); 5924 } 5925 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom(byte[] data) 5926 throws com.google.protobuf.InvalidProtocolBufferException { 5927 return PARSER.parseFrom(data); 5928 } 5929 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5930 byte[] data, 5931 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5932 throws com.google.protobuf.InvalidProtocolBufferException { 5933 return PARSER.parseFrom(data, extensionRegistry); 5934 } 5935 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom(java.io.InputStream input) 5936 throws java.io.IOException { 5937 return PARSER.parseFrom(input); 5938 } 5939 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5940 java.io.InputStream input, 5941 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5942 throws java.io.IOException { 5943 return PARSER.parseFrom(input, extensionRegistry); 5944 } 5945 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseDelimitedFrom(java.io.InputStream input) 5946 throws java.io.IOException { 5947 return PARSER.parseDelimitedFrom(input); 5948 } 5949 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseDelimitedFrom( 5950 java.io.InputStream input, 5951 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5952 throws java.io.IOException { 5953 return PARSER.parseDelimitedFrom(input, extensionRegistry); 5954 } 5955 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5956 com.google.protobuf.CodedInputStream input) 5957 throws java.io.IOException { 5958 return PARSER.parseFrom(input); 5959 } 5960 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 5961 com.google.protobuf.CodedInputStream input, 5962 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5963 throws java.io.IOException { 5964 return PARSER.parseFrom(input, extensionRegistry); 5965 } 5966 5967 public static Builder newBuilder() { return Builder.create(); } 5968 public Builder newBuilderForType() { return newBuilder(); } 5969 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory prototype) { 5970 return newBuilder().mergeFrom(prototype); 5971 } 5972 public Builder toBuilder() { return newBuilder(this); } 5973 5974 @java.lang.Override 5975 protected Builder newBuilderForType( 5976 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5977 Builder builder = new Builder(parent); 5978 return builder; 5979 } 5980 /** 5981 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeDirectory} 5982 */ 5983 public static final class Builder extends 5984 com.google.protobuf.GeneratedMessage.Builder<Builder> 5985 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder { 5986 public static final com.google.protobuf.Descriptors.Descriptor 5987 getDescriptor() { 5988 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 5989 } 5990 5991 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5992 internalGetFieldAccessorTable() { 5993 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable 5994 .ensureFieldAccessorsInitialized( 5995 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder.class); 5996 } 5997 5998 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder() 5999 private Builder() { 6000 maybeForceBuilderInitialization(); 6001 } 6002 6003 private Builder( 6004 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6005 super(parent); 6006 maybeForceBuilderInitialization(); 6007 } 6008 private void maybeForceBuilderInitialization() { 6009 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6010 getAclFieldBuilder(); 6011 } 6012 } 6013 private static Builder create() { 6014 return new Builder(); 6015 } 6016 6017 public Builder clear() { 6018 super.clear(); 6019 modificationTime_ = 0L; 6020 bitField0_ = (bitField0_ & ~0x00000001); 6021 nsQuota_ = 0L; 6022 bitField0_ = (bitField0_ & ~0x00000002); 6023 dsQuota_ = 0L; 6024 bitField0_ = (bitField0_ & ~0x00000004); 6025 permission_ = 0L; 6026 bitField0_ = (bitField0_ & ~0x00000008); 6027 if (aclBuilder_ == null) { 6028 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6029 } else { 6030 aclBuilder_.clear(); 6031 } 6032 bitField0_ = (bitField0_ & ~0x00000010); 6033 return this; 6034 } 6035 6036 public Builder clone() { 6037 return create().mergeFrom(buildPartial()); 6038 } 6039 6040 public com.google.protobuf.Descriptors.Descriptor 6041 getDescriptorForType() { 6042 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 6043 } 6044 6045 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDefaultInstanceForType() { 6046 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 6047 } 6048 6049 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory build() { 6050 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory result = buildPartial(); 6051 if (!result.isInitialized()) { 6052 throw newUninitializedMessageException(result); 6053 } 6054 return result; 6055 } 6056 6057 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory buildPartial() { 6058 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory(this); 6059 int from_bitField0_ = bitField0_; 6060 int to_bitField0_ = 0; 6061 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 6062 to_bitField0_ |= 0x00000001; 6063 } 6064 result.modificationTime_ = modificationTime_; 6065 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 6066 to_bitField0_ |= 0x00000002; 6067 } 6068 result.nsQuota_ = nsQuota_; 6069 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 6070 to_bitField0_ |= 0x00000004; 6071 } 6072 result.dsQuota_ = dsQuota_; 6073 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 6074 to_bitField0_ |= 0x00000008; 6075 } 6076 result.permission_ = permission_; 6077 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 6078 to_bitField0_ |= 0x00000010; 6079 } 6080 if (aclBuilder_ == null) { 6081 result.acl_ = acl_; 6082 } else { 6083 result.acl_ = aclBuilder_.build(); 6084 } 6085 result.bitField0_ = to_bitField0_; 6086 onBuilt(); 6087 return result; 6088 } 6089 6090 public Builder mergeFrom(com.google.protobuf.Message other) { 6091 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory) { 6092 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory)other); 6093 } else { 6094 super.mergeFrom(other); 6095 return this; 6096 } 6097 } 6098 6099 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory other) { 6100 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) return this; 6101 if (other.hasModificationTime()) { 6102 setModificationTime(other.getModificationTime()); 6103 } 6104 if (other.hasNsQuota()) { 6105 setNsQuota(other.getNsQuota()); 6106 } 6107 if (other.hasDsQuota()) { 6108 setDsQuota(other.getDsQuota()); 6109 } 6110 if (other.hasPermission()) { 6111 setPermission(other.getPermission()); 6112 } 6113 if (other.hasAcl()) { 6114 mergeAcl(other.getAcl()); 6115 } 6116 this.mergeUnknownFields(other.getUnknownFields()); 6117 return this; 6118 } 6119 6120 public final boolean isInitialized() { 6121 return true; 6122 } 6123 6124 public Builder mergeFrom( 6125 com.google.protobuf.CodedInputStream input, 6126 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6127 throws java.io.IOException { 6128 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parsedMessage = null; 6129 try { 6130 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6131 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6132 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory) e.getUnfinishedMessage(); 6133 throw e; 6134 } finally { 6135 if (parsedMessage != null) { 6136 mergeFrom(parsedMessage); 6137 } 6138 } 6139 return this; 6140 } 6141 private int bitField0_; 6142 6143 // optional uint64 modificationTime = 1; 6144 private long modificationTime_ ; 6145 /** 6146 * <code>optional uint64 modificationTime = 1;</code> 6147 */ 6148 public boolean hasModificationTime() { 6149 return ((bitField0_ & 0x00000001) == 0x00000001); 6150 } 6151 /** 6152 * <code>optional uint64 modificationTime = 1;</code> 6153 */ 6154 public long getModificationTime() { 6155 return modificationTime_; 6156 } 6157 /** 6158 * <code>optional uint64 modificationTime = 1;</code> 6159 */ 6160 public Builder setModificationTime(long value) { 6161 bitField0_ |= 0x00000001; 6162 modificationTime_ = value; 6163 onChanged(); 6164 return this; 6165 } 6166 /** 6167 * <code>optional uint64 modificationTime = 1;</code> 6168 */ 6169 public Builder clearModificationTime() { 6170 bitField0_ = (bitField0_ & ~0x00000001); 6171 modificationTime_ = 0L; 6172 onChanged(); 6173 return this; 6174 } 6175 6176 // optional uint64 nsQuota = 2; 6177 private long nsQuota_ ; 6178 /** 6179 * <code>optional uint64 nsQuota = 2;</code> 6180 * 6181 * <pre> 6182 * namespace quota 6183 * </pre> 6184 */ 6185 public boolean hasNsQuota() { 6186 return ((bitField0_ & 0x00000002) == 0x00000002); 6187 } 6188 /** 6189 * <code>optional uint64 nsQuota = 2;</code> 6190 * 6191 * <pre> 6192 * namespace quota 6193 * </pre> 6194 */ 6195 public long getNsQuota() { 6196 return nsQuota_; 6197 } 6198 /** 6199 * <code>optional uint64 nsQuota = 2;</code> 6200 * 6201 * <pre> 6202 * namespace quota 6203 * </pre> 6204 */ 6205 public Builder setNsQuota(long value) { 6206 bitField0_ |= 0x00000002; 6207 nsQuota_ = value; 6208 onChanged(); 6209 return this; 6210 } 6211 /** 6212 * <code>optional uint64 nsQuota = 2;</code> 6213 * 6214 * <pre> 6215 * namespace quota 6216 * </pre> 6217 */ 6218 public Builder clearNsQuota() { 6219 bitField0_ = (bitField0_ & ~0x00000002); 6220 nsQuota_ = 0L; 6221 onChanged(); 6222 return this; 6223 } 6224 6225 // optional uint64 dsQuota = 3; 6226 private long dsQuota_ ; 6227 /** 6228 * <code>optional uint64 dsQuota = 3;</code> 6229 * 6230 * <pre> 6231 * diskspace quota 6232 * </pre> 6233 */ 6234 public boolean hasDsQuota() { 6235 return ((bitField0_ & 0x00000004) == 0x00000004); 6236 } 6237 /** 6238 * <code>optional uint64 dsQuota = 3;</code> 6239 * 6240 * <pre> 6241 * diskspace quota 6242 * </pre> 6243 */ 6244 public long getDsQuota() { 6245 return dsQuota_; 6246 } 6247 /** 6248 * <code>optional uint64 dsQuota = 3;</code> 6249 * 6250 * <pre> 6251 * diskspace quota 6252 * </pre> 6253 */ 6254 public Builder setDsQuota(long value) { 6255 bitField0_ |= 0x00000004; 6256 dsQuota_ = value; 6257 onChanged(); 6258 return this; 6259 } 6260 /** 6261 * <code>optional uint64 dsQuota = 3;</code> 6262 * 6263 * <pre> 6264 * diskspace quota 6265 * </pre> 6266 */ 6267 public Builder clearDsQuota() { 6268 bitField0_ = (bitField0_ & ~0x00000004); 6269 dsQuota_ = 0L; 6270 onChanged(); 6271 return this; 6272 } 6273 6274 // optional fixed64 permission = 4; 6275 private long permission_ ; 6276 /** 6277 * <code>optional fixed64 permission = 4;</code> 6278 */ 6279 public boolean hasPermission() { 6280 return ((bitField0_ & 0x00000008) == 0x00000008); 6281 } 6282 /** 6283 * <code>optional fixed64 permission = 4;</code> 6284 */ 6285 public long getPermission() { 6286 return permission_; 6287 } 6288 /** 6289 * <code>optional fixed64 permission = 4;</code> 6290 */ 6291 public Builder setPermission(long value) { 6292 bitField0_ |= 0x00000008; 6293 permission_ = value; 6294 onChanged(); 6295 return this; 6296 } 6297 /** 6298 * <code>optional fixed64 permission = 4;</code> 6299 */ 6300 public Builder clearPermission() { 6301 bitField0_ = (bitField0_ & ~0x00000008); 6302 permission_ = 0L; 6303 onChanged(); 6304 return this; 6305 } 6306 6307 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 6308 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6309 private com.google.protobuf.SingleFieldBuilder< 6310 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> aclBuilder_; 6311 /** 6312 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6313 */ 6314 public boolean hasAcl() { 6315 return ((bitField0_ & 0x00000010) == 0x00000010); 6316 } 6317 /** 6318 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6319 */ 6320 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 6321 if (aclBuilder_ == null) { 6322 return acl_; 6323 } else { 6324 return aclBuilder_.getMessage(); 6325 } 6326 } 6327 /** 6328 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6329 */ 6330 public Builder setAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 6331 if (aclBuilder_ == null) { 6332 if (value == null) { 6333 throw new NullPointerException(); 6334 } 6335 acl_ = value; 6336 onChanged(); 6337 } else { 6338 aclBuilder_.setMessage(value); 6339 } 6340 bitField0_ |= 0x00000010; 6341 return this; 6342 } 6343 /** 6344 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6345 */ 6346 public Builder setAcl( 6347 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder builderForValue) { 6348 if (aclBuilder_ == null) { 6349 acl_ = builderForValue.build(); 6350 onChanged(); 6351 } else { 6352 aclBuilder_.setMessage(builderForValue.build()); 6353 } 6354 bitField0_ |= 0x00000010; 6355 return this; 6356 } 6357 /** 6358 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6359 */ 6360 public Builder mergeAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 6361 if (aclBuilder_ == null) { 6362 if (((bitField0_ & 0x00000010) == 0x00000010) && 6363 acl_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) { 6364 acl_ = 6365 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder(acl_).mergeFrom(value).buildPartial(); 6366 } else { 6367 acl_ = value; 6368 } 6369 onChanged(); 6370 } else { 6371 aclBuilder_.mergeFrom(value); 6372 } 6373 bitField0_ |= 0x00000010; 6374 return this; 6375 } 6376 /** 6377 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6378 */ 6379 public Builder clearAcl() { 6380 if (aclBuilder_ == null) { 6381 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6382 onChanged(); 6383 } else { 6384 aclBuilder_.clear(); 6385 } 6386 bitField0_ = (bitField0_ & ~0x00000010); 6387 return this; 6388 } 6389 /** 6390 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6391 */ 6392 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder getAclBuilder() { 6393 bitField0_ |= 0x00000010; 6394 onChanged(); 6395 return getAclFieldBuilder().getBuilder(); 6396 } 6397 /** 6398 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6399 */ 6400 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 6401 if (aclBuilder_ != null) { 6402 return aclBuilder_.getMessageOrBuilder(); 6403 } else { 6404 return acl_; 6405 } 6406 } 6407 /** 6408 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 6409 */ 6410 private com.google.protobuf.SingleFieldBuilder< 6411 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> 6412 getAclFieldBuilder() { 6413 if (aclBuilder_ == null) { 6414 aclBuilder_ = new com.google.protobuf.SingleFieldBuilder< 6415 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder>( 6416 acl_, 6417 getParentForChildren(), 6418 isClean()); 6419 acl_ = null; 6420 } 6421 return aclBuilder_; 6422 } 6423 6424 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeDirectory) 6425 } 6426 6427 static { 6428 defaultInstance = new INodeDirectory(true); 6429 defaultInstance.initFields(); 6430 } 6431 6432 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeDirectory) 6433 } 6434 6435 public interface INodeSymlinkOrBuilder 6436 extends com.google.protobuf.MessageOrBuilder { 6437 6438 // optional fixed64 permission = 1; 6439 /** 6440 * <code>optional fixed64 permission = 1;</code> 6441 */ 6442 boolean hasPermission(); 6443 /** 6444 * <code>optional fixed64 permission = 1;</code> 6445 */ 6446 long getPermission(); 6447 6448 // optional bytes target = 2; 6449 /** 6450 * <code>optional bytes target = 2;</code> 6451 */ 6452 boolean hasTarget(); 6453 /** 6454 * <code>optional bytes target = 2;</code> 6455 */ 6456 com.google.protobuf.ByteString getTarget(); 6457 6458 // optional uint64 modificationTime = 3; 6459 /** 6460 * <code>optional uint64 modificationTime = 3;</code> 6461 */ 6462 boolean hasModificationTime(); 6463 /** 6464 * <code>optional uint64 modificationTime = 3;</code> 6465 */ 6466 long getModificationTime(); 6467 6468 // optional uint64 accessTime = 4; 6469 /** 6470 * <code>optional uint64 accessTime = 4;</code> 6471 */ 6472 boolean hasAccessTime(); 6473 /** 6474 * <code>optional uint64 accessTime = 4;</code> 6475 */ 6476 long getAccessTime(); 6477 } 6478 /** 6479 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeSymlink} 6480 */ 6481 public static final class INodeSymlink extends 6482 com.google.protobuf.GeneratedMessage 6483 implements INodeSymlinkOrBuilder { 6484 // Use INodeSymlink.newBuilder() to construct. 6485 private INodeSymlink(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 6486 super(builder); 6487 this.unknownFields = builder.getUnknownFields(); 6488 } 6489 private INodeSymlink(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 6490 6491 private static final INodeSymlink defaultInstance; 6492 public static INodeSymlink getDefaultInstance() { 6493 return defaultInstance; 6494 } 6495 6496 public INodeSymlink getDefaultInstanceForType() { 6497 return defaultInstance; 6498 } 6499 6500 private final com.google.protobuf.UnknownFieldSet unknownFields; 6501 @java.lang.Override 6502 public final com.google.protobuf.UnknownFieldSet 6503 getUnknownFields() { 6504 return this.unknownFields; 6505 } 6506 private INodeSymlink( 6507 com.google.protobuf.CodedInputStream input, 6508 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6509 throws com.google.protobuf.InvalidProtocolBufferException { 6510 initFields(); 6511 int mutable_bitField0_ = 0; 6512 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 6513 com.google.protobuf.UnknownFieldSet.newBuilder(); 6514 try { 6515 boolean done = false; 6516 while (!done) { 6517 int tag = input.readTag(); 6518 switch (tag) { 6519 case 0: 6520 done = true; 6521 break; 6522 default: { 6523 if (!parseUnknownField(input, unknownFields, 6524 extensionRegistry, tag)) { 6525 done = true; 6526 } 6527 break; 6528 } 6529 case 9: { 6530 bitField0_ |= 0x00000001; 6531 permission_ = input.readFixed64(); 6532 break; 6533 } 6534 case 18: { 6535 bitField0_ |= 0x00000002; 6536 target_ = input.readBytes(); 6537 break; 6538 } 6539 case 24: { 6540 bitField0_ |= 0x00000004; 6541 modificationTime_ = input.readUInt64(); 6542 break; 6543 } 6544 case 32: { 6545 bitField0_ |= 0x00000008; 6546 accessTime_ = input.readUInt64(); 6547 break; 6548 } 6549 } 6550 } 6551 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6552 throw e.setUnfinishedMessage(this); 6553 } catch (java.io.IOException e) { 6554 throw new com.google.protobuf.InvalidProtocolBufferException( 6555 e.getMessage()).setUnfinishedMessage(this); 6556 } finally { 6557 this.unknownFields = unknownFields.build(); 6558 makeExtensionsImmutable(); 6559 } 6560 } 6561 public static final com.google.protobuf.Descriptors.Descriptor 6562 getDescriptor() { 6563 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 6564 } 6565 6566 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6567 internalGetFieldAccessorTable() { 6568 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable 6569 .ensureFieldAccessorsInitialized( 6570 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder.class); 6571 } 6572 6573 public static com.google.protobuf.Parser<INodeSymlink> PARSER = 6574 new com.google.protobuf.AbstractParser<INodeSymlink>() { 6575 public INodeSymlink parsePartialFrom( 6576 com.google.protobuf.CodedInputStream input, 6577 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6578 throws com.google.protobuf.InvalidProtocolBufferException { 6579 return new INodeSymlink(input, extensionRegistry); 6580 } 6581 }; 6582 6583 @java.lang.Override 6584 public com.google.protobuf.Parser<INodeSymlink> getParserForType() { 6585 return PARSER; 6586 } 6587 6588 private int bitField0_; 6589 // optional fixed64 permission = 1; 6590 public static final int PERMISSION_FIELD_NUMBER = 1; 6591 private long permission_; 6592 /** 6593 * <code>optional fixed64 permission = 1;</code> 6594 */ 6595 public boolean hasPermission() { 6596 return ((bitField0_ & 0x00000001) == 0x00000001); 6597 } 6598 /** 6599 * <code>optional fixed64 permission = 1;</code> 6600 */ 6601 public long getPermission() { 6602 return permission_; 6603 } 6604 6605 // optional bytes target = 2; 6606 public static final int TARGET_FIELD_NUMBER = 2; 6607 private com.google.protobuf.ByteString target_; 6608 /** 6609 * <code>optional bytes target = 2;</code> 6610 */ 6611 public boolean hasTarget() { 6612 return ((bitField0_ & 0x00000002) == 0x00000002); 6613 } 6614 /** 6615 * <code>optional bytes target = 2;</code> 6616 */ 6617 public com.google.protobuf.ByteString getTarget() { 6618 return target_; 6619 } 6620 6621 // optional uint64 modificationTime = 3; 6622 public static final int MODIFICATIONTIME_FIELD_NUMBER = 3; 6623 private long modificationTime_; 6624 /** 6625 * <code>optional uint64 modificationTime = 3;</code> 6626 */ 6627 public boolean hasModificationTime() { 6628 return ((bitField0_ & 0x00000004) == 0x00000004); 6629 } 6630 /** 6631 * <code>optional uint64 modificationTime = 3;</code> 6632 */ 6633 public long getModificationTime() { 6634 return modificationTime_; 6635 } 6636 6637 // optional uint64 accessTime = 4; 6638 public static final int ACCESSTIME_FIELD_NUMBER = 4; 6639 private long accessTime_; 6640 /** 6641 * <code>optional uint64 accessTime = 4;</code> 6642 */ 6643 public boolean hasAccessTime() { 6644 return ((bitField0_ & 0x00000008) == 0x00000008); 6645 } 6646 /** 6647 * <code>optional uint64 accessTime = 4;</code> 6648 */ 6649 public long getAccessTime() { 6650 return accessTime_; 6651 } 6652 6653 private void initFields() { 6654 permission_ = 0L; 6655 target_ = com.google.protobuf.ByteString.EMPTY; 6656 modificationTime_ = 0L; 6657 accessTime_ = 0L; 6658 } 6659 private byte memoizedIsInitialized = -1; 6660 public final boolean isInitialized() { 6661 byte isInitialized = memoizedIsInitialized; 6662 if (isInitialized != -1) return isInitialized == 1; 6663 6664 memoizedIsInitialized = 1; 6665 return true; 6666 } 6667 6668 public void writeTo(com.google.protobuf.CodedOutputStream output) 6669 throws java.io.IOException { 6670 getSerializedSize(); 6671 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6672 output.writeFixed64(1, permission_); 6673 } 6674 if (((bitField0_ & 0x00000002) == 0x00000002)) { 6675 output.writeBytes(2, target_); 6676 } 6677 if (((bitField0_ & 0x00000004) == 0x00000004)) { 6678 output.writeUInt64(3, modificationTime_); 6679 } 6680 if (((bitField0_ & 0x00000008) == 0x00000008)) { 6681 output.writeUInt64(4, accessTime_); 6682 } 6683 getUnknownFields().writeTo(output); 6684 } 6685 6686 private int memoizedSerializedSize = -1; 6687 public int getSerializedSize() { 6688 int size = memoizedSerializedSize; 6689 if (size != -1) return size; 6690 6691 size = 0; 6692 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6693 size += com.google.protobuf.CodedOutputStream 6694 .computeFixed64Size(1, permission_); 6695 } 6696 if (((bitField0_ & 0x00000002) == 0x00000002)) { 6697 size += com.google.protobuf.CodedOutputStream 6698 .computeBytesSize(2, target_); 6699 } 6700 if (((bitField0_ & 0x00000004) == 0x00000004)) { 6701 size += com.google.protobuf.CodedOutputStream 6702 .computeUInt64Size(3, modificationTime_); 6703 } 6704 if (((bitField0_ & 0x00000008) == 0x00000008)) { 6705 size += com.google.protobuf.CodedOutputStream 6706 .computeUInt64Size(4, accessTime_); 6707 } 6708 size += getUnknownFields().getSerializedSize(); 6709 memoizedSerializedSize = size; 6710 return size; 6711 } 6712 6713 private static final long serialVersionUID = 0L; 6714 @java.lang.Override 6715 protected java.lang.Object writeReplace() 6716 throws java.io.ObjectStreamException { 6717 return super.writeReplace(); 6718 } 6719 6720 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6721 com.google.protobuf.ByteString data) 6722 throws com.google.protobuf.InvalidProtocolBufferException { 6723 return PARSER.parseFrom(data); 6724 } 6725 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6726 com.google.protobuf.ByteString data, 6727 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6728 throws com.google.protobuf.InvalidProtocolBufferException { 6729 return PARSER.parseFrom(data, extensionRegistry); 6730 } 6731 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom(byte[] data) 6732 throws com.google.protobuf.InvalidProtocolBufferException { 6733 return PARSER.parseFrom(data); 6734 } 6735 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6736 byte[] data, 6737 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6738 throws com.google.protobuf.InvalidProtocolBufferException { 6739 return PARSER.parseFrom(data, extensionRegistry); 6740 } 6741 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom(java.io.InputStream input) 6742 throws java.io.IOException { 6743 return PARSER.parseFrom(input); 6744 } 6745 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6746 java.io.InputStream input, 6747 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6748 throws java.io.IOException { 6749 return PARSER.parseFrom(input, extensionRegistry); 6750 } 6751 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseDelimitedFrom(java.io.InputStream input) 6752 throws java.io.IOException { 6753 return PARSER.parseDelimitedFrom(input); 6754 } 6755 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseDelimitedFrom( 6756 java.io.InputStream input, 6757 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6758 throws java.io.IOException { 6759 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6760 } 6761 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6762 com.google.protobuf.CodedInputStream input) 6763 throws java.io.IOException { 6764 return PARSER.parseFrom(input); 6765 } 6766 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 6767 com.google.protobuf.CodedInputStream input, 6768 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6769 throws java.io.IOException { 6770 return PARSER.parseFrom(input, extensionRegistry); 6771 } 6772 6773 public static Builder newBuilder() { return Builder.create(); } 6774 public Builder newBuilderForType() { return newBuilder(); } 6775 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink prototype) { 6776 return newBuilder().mergeFrom(prototype); 6777 } 6778 public Builder toBuilder() { return newBuilder(this); } 6779 6780 @java.lang.Override 6781 protected Builder newBuilderForType( 6782 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6783 Builder builder = new Builder(parent); 6784 return builder; 6785 } 6786 /** 6787 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeSymlink} 6788 */ 6789 public static final class Builder extends 6790 com.google.protobuf.GeneratedMessage.Builder<Builder> 6791 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder { 6792 public static final com.google.protobuf.Descriptors.Descriptor 6793 getDescriptor() { 6794 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 6795 } 6796 6797 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6798 internalGetFieldAccessorTable() { 6799 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable 6800 .ensureFieldAccessorsInitialized( 6801 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder.class); 6802 } 6803 6804 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.newBuilder() 6805 private Builder() { 6806 maybeForceBuilderInitialization(); 6807 } 6808 6809 private Builder( 6810 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6811 super(parent); 6812 maybeForceBuilderInitialization(); 6813 } 6814 private void maybeForceBuilderInitialization() { 6815 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6816 } 6817 } 6818 private static Builder create() { 6819 return new Builder(); 6820 } 6821 6822 public Builder clear() { 6823 super.clear(); 6824 permission_ = 0L; 6825 bitField0_ = (bitField0_ & ~0x00000001); 6826 target_ = com.google.protobuf.ByteString.EMPTY; 6827 bitField0_ = (bitField0_ & ~0x00000002); 6828 modificationTime_ = 0L; 6829 bitField0_ = (bitField0_ & ~0x00000004); 6830 accessTime_ = 0L; 6831 bitField0_ = (bitField0_ & ~0x00000008); 6832 return this; 6833 } 6834 6835 public Builder clone() { 6836 return create().mergeFrom(buildPartial()); 6837 } 6838 6839 public com.google.protobuf.Descriptors.Descriptor 6840 getDescriptorForType() { 6841 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 6842 } 6843 6844 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getDefaultInstanceForType() { 6845 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 6846 } 6847 6848 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink build() { 6849 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink result = buildPartial(); 6850 if (!result.isInitialized()) { 6851 throw newUninitializedMessageException(result); 6852 } 6853 return result; 6854 } 6855 6856 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink buildPartial() { 6857 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink(this); 6858 int from_bitField0_ = bitField0_; 6859 int to_bitField0_ = 0; 6860 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 6861 to_bitField0_ |= 0x00000001; 6862 } 6863 result.permission_ = permission_; 6864 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 6865 to_bitField0_ |= 0x00000002; 6866 } 6867 result.target_ = target_; 6868 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 6869 to_bitField0_ |= 0x00000004; 6870 } 6871 result.modificationTime_ = modificationTime_; 6872 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 6873 to_bitField0_ |= 0x00000008; 6874 } 6875 result.accessTime_ = accessTime_; 6876 result.bitField0_ = to_bitField0_; 6877 onBuilt(); 6878 return result; 6879 } 6880 6881 public Builder mergeFrom(com.google.protobuf.Message other) { 6882 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink) { 6883 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink)other); 6884 } else { 6885 super.mergeFrom(other); 6886 return this; 6887 } 6888 } 6889 6890 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink other) { 6891 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance()) return this; 6892 if (other.hasPermission()) { 6893 setPermission(other.getPermission()); 6894 } 6895 if (other.hasTarget()) { 6896 setTarget(other.getTarget()); 6897 } 6898 if (other.hasModificationTime()) { 6899 setModificationTime(other.getModificationTime()); 6900 } 6901 if (other.hasAccessTime()) { 6902 setAccessTime(other.getAccessTime()); 6903 } 6904 this.mergeUnknownFields(other.getUnknownFields()); 6905 return this; 6906 } 6907 6908 public final boolean isInitialized() { 6909 return true; 6910 } 6911 6912 public Builder mergeFrom( 6913 com.google.protobuf.CodedInputStream input, 6914 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6915 throws java.io.IOException { 6916 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parsedMessage = null; 6917 try { 6918 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6919 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6920 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink) e.getUnfinishedMessage(); 6921 throw e; 6922 } finally { 6923 if (parsedMessage != null) { 6924 mergeFrom(parsedMessage); 6925 } 6926 } 6927 return this; 6928 } 6929 private int bitField0_; 6930 6931 // optional fixed64 permission = 1; 6932 private long permission_ ; 6933 /** 6934 * <code>optional fixed64 permission = 1;</code> 6935 */ 6936 public boolean hasPermission() { 6937 return ((bitField0_ & 0x00000001) == 0x00000001); 6938 } 6939 /** 6940 * <code>optional fixed64 permission = 1;</code> 6941 */ 6942 public long getPermission() { 6943 return permission_; 6944 } 6945 /** 6946 * <code>optional fixed64 permission = 1;</code> 6947 */ 6948 public Builder setPermission(long value) { 6949 bitField0_ |= 0x00000001; 6950 permission_ = value; 6951 onChanged(); 6952 return this; 6953 } 6954 /** 6955 * <code>optional fixed64 permission = 1;</code> 6956 */ 6957 public Builder clearPermission() { 6958 bitField0_ = (bitField0_ & ~0x00000001); 6959 permission_ = 0L; 6960 onChanged(); 6961 return this; 6962 } 6963 6964 // optional bytes target = 2; 6965 private com.google.protobuf.ByteString target_ = com.google.protobuf.ByteString.EMPTY; 6966 /** 6967 * <code>optional bytes target = 2;</code> 6968 */ 6969 public boolean hasTarget() { 6970 return ((bitField0_ & 0x00000002) == 0x00000002); 6971 } 6972 /** 6973 * <code>optional bytes target = 2;</code> 6974 */ 6975 public com.google.protobuf.ByteString getTarget() { 6976 return target_; 6977 } 6978 /** 6979 * <code>optional bytes target = 2;</code> 6980 */ 6981 public Builder setTarget(com.google.protobuf.ByteString value) { 6982 if (value == null) { 6983 throw new NullPointerException(); 6984 } 6985 bitField0_ |= 0x00000002; 6986 target_ = value; 6987 onChanged(); 6988 return this; 6989 } 6990 /** 6991 * <code>optional bytes target = 2;</code> 6992 */ 6993 public Builder clearTarget() { 6994 bitField0_ = (bitField0_ & ~0x00000002); 6995 target_ = getDefaultInstance().getTarget(); 6996 onChanged(); 6997 return this; 6998 } 6999 7000 // optional uint64 modificationTime = 3; 7001 private long modificationTime_ ; 7002 /** 7003 * <code>optional uint64 modificationTime = 3;</code> 7004 */ 7005 public boolean hasModificationTime() { 7006 return ((bitField0_ & 0x00000004) == 0x00000004); 7007 } 7008 /** 7009 * <code>optional uint64 modificationTime = 3;</code> 7010 */ 7011 public long getModificationTime() { 7012 return modificationTime_; 7013 } 7014 /** 7015 * <code>optional uint64 modificationTime = 3;</code> 7016 */ 7017 public Builder setModificationTime(long value) { 7018 bitField0_ |= 0x00000004; 7019 modificationTime_ = value; 7020 onChanged(); 7021 return this; 7022 } 7023 /** 7024 * <code>optional uint64 modificationTime = 3;</code> 7025 */ 7026 public Builder clearModificationTime() { 7027 bitField0_ = (bitField0_ & ~0x00000004); 7028 modificationTime_ = 0L; 7029 onChanged(); 7030 return this; 7031 } 7032 7033 // optional uint64 accessTime = 4; 7034 private long accessTime_ ; 7035 /** 7036 * <code>optional uint64 accessTime = 4;</code> 7037 */ 7038 public boolean hasAccessTime() { 7039 return ((bitField0_ & 0x00000008) == 0x00000008); 7040 } 7041 /** 7042 * <code>optional uint64 accessTime = 4;</code> 7043 */ 7044 public long getAccessTime() { 7045 return accessTime_; 7046 } 7047 /** 7048 * <code>optional uint64 accessTime = 4;</code> 7049 */ 7050 public Builder setAccessTime(long value) { 7051 bitField0_ |= 0x00000008; 7052 accessTime_ = value; 7053 onChanged(); 7054 return this; 7055 } 7056 /** 7057 * <code>optional uint64 accessTime = 4;</code> 7058 */ 7059 public Builder clearAccessTime() { 7060 bitField0_ = (bitField0_ & ~0x00000008); 7061 accessTime_ = 0L; 7062 onChanged(); 7063 return this; 7064 } 7065 7066 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeSymlink) 7067 } 7068 7069 static { 7070 defaultInstance = new INodeSymlink(true); 7071 defaultInstance.initFields(); 7072 } 7073 7074 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeSymlink) 7075 } 7076 7077 public interface INodeOrBuilder 7078 extends com.google.protobuf.MessageOrBuilder { 7079 7080 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 7081 /** 7082 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7083 */ 7084 boolean hasType(); 7085 /** 7086 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7087 */ 7088 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType(); 7089 7090 // required uint64 id = 2; 7091 /** 7092 * <code>required uint64 id = 2;</code> 7093 */ 7094 boolean hasId(); 7095 /** 7096 * <code>required uint64 id = 2;</code> 7097 */ 7098 long getId(); 7099 7100 // optional bytes name = 3; 7101 /** 7102 * <code>optional bytes name = 3;</code> 7103 */ 7104 boolean hasName(); 7105 /** 7106 * <code>optional bytes name = 3;</code> 7107 */ 7108 com.google.protobuf.ByteString getName(); 7109 7110 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 7111 /** 7112 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7113 */ 7114 boolean hasFile(); 7115 /** 7116 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7117 */ 7118 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile(); 7119 /** 7120 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7121 */ 7122 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder(); 7123 7124 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 7125 /** 7126 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7127 */ 7128 boolean hasDirectory(); 7129 /** 7130 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7131 */ 7132 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory(); 7133 /** 7134 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7135 */ 7136 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder(); 7137 7138 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 7139 /** 7140 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7141 */ 7142 boolean hasSymlink(); 7143 /** 7144 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7145 */ 7146 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink(); 7147 /** 7148 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7149 */ 7150 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder(); 7151 } 7152 /** 7153 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INode} 7154 */ 7155 public static final class INode extends 7156 com.google.protobuf.GeneratedMessage 7157 implements INodeOrBuilder { 7158 // Use INode.newBuilder() to construct. 7159 private INode(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7160 super(builder); 7161 this.unknownFields = builder.getUnknownFields(); 7162 } 7163 private INode(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7164 7165 private static final INode defaultInstance; 7166 public static INode getDefaultInstance() { 7167 return defaultInstance; 7168 } 7169 7170 public INode getDefaultInstanceForType() { 7171 return defaultInstance; 7172 } 7173 7174 private final com.google.protobuf.UnknownFieldSet unknownFields; 7175 @java.lang.Override 7176 public final com.google.protobuf.UnknownFieldSet 7177 getUnknownFields() { 7178 return this.unknownFields; 7179 } 7180 private INode( 7181 com.google.protobuf.CodedInputStream input, 7182 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7183 throws com.google.protobuf.InvalidProtocolBufferException { 7184 initFields(); 7185 int mutable_bitField0_ = 0; 7186 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7187 com.google.protobuf.UnknownFieldSet.newBuilder(); 7188 try { 7189 boolean done = false; 7190 while (!done) { 7191 int tag = input.readTag(); 7192 switch (tag) { 7193 case 0: 7194 done = true; 7195 break; 7196 default: { 7197 if (!parseUnknownField(input, unknownFields, 7198 extensionRegistry, tag)) { 7199 done = true; 7200 } 7201 break; 7202 } 7203 case 8: { 7204 int rawValue = input.readEnum(); 7205 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type value = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.valueOf(rawValue); 7206 if (value == null) { 7207 unknownFields.mergeVarintField(1, rawValue); 7208 } else { 7209 bitField0_ |= 0x00000001; 7210 type_ = value; 7211 } 7212 break; 7213 } 7214 case 16: { 7215 bitField0_ |= 0x00000002; 7216 id_ = input.readUInt64(); 7217 break; 7218 } 7219 case 26: { 7220 bitField0_ |= 0x00000004; 7221 name_ = input.readBytes(); 7222 break; 7223 } 7224 case 34: { 7225 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder subBuilder = null; 7226 if (((bitField0_ & 0x00000008) == 0x00000008)) { 7227 subBuilder = file_.toBuilder(); 7228 } 7229 file_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.PARSER, extensionRegistry); 7230 if (subBuilder != null) { 7231 subBuilder.mergeFrom(file_); 7232 file_ = subBuilder.buildPartial(); 7233 } 7234 bitField0_ |= 0x00000008; 7235 break; 7236 } 7237 case 42: { 7238 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder subBuilder = null; 7239 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7240 subBuilder = directory_.toBuilder(); 7241 } 7242 directory_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.PARSER, extensionRegistry); 7243 if (subBuilder != null) { 7244 subBuilder.mergeFrom(directory_); 7245 directory_ = subBuilder.buildPartial(); 7246 } 7247 bitField0_ |= 0x00000010; 7248 break; 7249 } 7250 case 50: { 7251 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder subBuilder = null; 7252 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7253 subBuilder = symlink_.toBuilder(); 7254 } 7255 symlink_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.PARSER, extensionRegistry); 7256 if (subBuilder != null) { 7257 subBuilder.mergeFrom(symlink_); 7258 symlink_ = subBuilder.buildPartial(); 7259 } 7260 bitField0_ |= 0x00000020; 7261 break; 7262 } 7263 } 7264 } 7265 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7266 throw e.setUnfinishedMessage(this); 7267 } catch (java.io.IOException e) { 7268 throw new com.google.protobuf.InvalidProtocolBufferException( 7269 e.getMessage()).setUnfinishedMessage(this); 7270 } finally { 7271 this.unknownFields = unknownFields.build(); 7272 makeExtensionsImmutable(); 7273 } 7274 } 7275 public static final com.google.protobuf.Descriptors.Descriptor 7276 getDescriptor() { 7277 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 7278 } 7279 7280 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7281 internalGetFieldAccessorTable() { 7282 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable 7283 .ensureFieldAccessorsInitialized( 7284 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder.class); 7285 } 7286 7287 public static com.google.protobuf.Parser<INode> PARSER = 7288 new com.google.protobuf.AbstractParser<INode>() { 7289 public INode parsePartialFrom( 7290 com.google.protobuf.CodedInputStream input, 7291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7292 throws com.google.protobuf.InvalidProtocolBufferException { 7293 return new INode(input, extensionRegistry); 7294 } 7295 }; 7296 7297 @java.lang.Override 7298 public com.google.protobuf.Parser<INode> getParserForType() { 7299 return PARSER; 7300 } 7301 7302 /** 7303 * Protobuf enum {@code hadoop.hdfs.fsimage.INodeSection.INode.Type} 7304 */ 7305 public enum Type 7306 implements com.google.protobuf.ProtocolMessageEnum { 7307 /** 7308 * <code>FILE = 1;</code> 7309 */ 7310 FILE(0, 1), 7311 /** 7312 * <code>DIRECTORY = 2;</code> 7313 */ 7314 DIRECTORY(1, 2), 7315 /** 7316 * <code>SYMLINK = 3;</code> 7317 */ 7318 SYMLINK(2, 3), 7319 ; 7320 7321 /** 7322 * <code>FILE = 1;</code> 7323 */ 7324 public static final int FILE_VALUE = 1; 7325 /** 7326 * <code>DIRECTORY = 2;</code> 7327 */ 7328 public static final int DIRECTORY_VALUE = 2; 7329 /** 7330 * <code>SYMLINK = 3;</code> 7331 */ 7332 public static final int SYMLINK_VALUE = 3; 7333 7334 7335 public final int getNumber() { return value; } 7336 7337 public static Type valueOf(int value) { 7338 switch (value) { 7339 case 1: return FILE; 7340 case 2: return DIRECTORY; 7341 case 3: return SYMLINK; 7342 default: return null; 7343 } 7344 } 7345 7346 public static com.google.protobuf.Internal.EnumLiteMap<Type> 7347 internalGetValueMap() { 7348 return internalValueMap; 7349 } 7350 private static com.google.protobuf.Internal.EnumLiteMap<Type> 7351 internalValueMap = 7352 new com.google.protobuf.Internal.EnumLiteMap<Type>() { 7353 public Type findValueByNumber(int number) { 7354 return Type.valueOf(number); 7355 } 7356 }; 7357 7358 public final com.google.protobuf.Descriptors.EnumValueDescriptor 7359 getValueDescriptor() { 7360 return getDescriptor().getValues().get(index); 7361 } 7362 public final com.google.protobuf.Descriptors.EnumDescriptor 7363 getDescriptorForType() { 7364 return getDescriptor(); 7365 } 7366 public static final com.google.protobuf.Descriptors.EnumDescriptor 7367 getDescriptor() { 7368 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDescriptor().getEnumTypes().get(0); 7369 } 7370 7371 private static final Type[] VALUES = values(); 7372 7373 public static Type valueOf( 7374 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 7375 if (desc.getType() != getDescriptor()) { 7376 throw new java.lang.IllegalArgumentException( 7377 "EnumValueDescriptor is not for this type."); 7378 } 7379 return VALUES[desc.getIndex()]; 7380 } 7381 7382 private final int index; 7383 private final int value; 7384 7385 private Type(int index, int value) { 7386 this.index = index; 7387 this.value = value; 7388 } 7389 7390 // @@protoc_insertion_point(enum_scope:hadoop.hdfs.fsimage.INodeSection.INode.Type) 7391 } 7392 7393 private int bitField0_; 7394 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 7395 public static final int TYPE_FIELD_NUMBER = 1; 7396 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type type_; 7397 /** 7398 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7399 */ 7400 public boolean hasType() { 7401 return ((bitField0_ & 0x00000001) == 0x00000001); 7402 } 7403 /** 7404 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7405 */ 7406 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType() { 7407 return type_; 7408 } 7409 7410 // required uint64 id = 2; 7411 public static final int ID_FIELD_NUMBER = 2; 7412 private long id_; 7413 /** 7414 * <code>required uint64 id = 2;</code> 7415 */ 7416 public boolean hasId() { 7417 return ((bitField0_ & 0x00000002) == 0x00000002); 7418 } 7419 /** 7420 * <code>required uint64 id = 2;</code> 7421 */ 7422 public long getId() { 7423 return id_; 7424 } 7425 7426 // optional bytes name = 3; 7427 public static final int NAME_FIELD_NUMBER = 3; 7428 private com.google.protobuf.ByteString name_; 7429 /** 7430 * <code>optional bytes name = 3;</code> 7431 */ 7432 public boolean hasName() { 7433 return ((bitField0_ & 0x00000004) == 0x00000004); 7434 } 7435 /** 7436 * <code>optional bytes name = 3;</code> 7437 */ 7438 public com.google.protobuf.ByteString getName() { 7439 return name_; 7440 } 7441 7442 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 7443 public static final int FILE_FIELD_NUMBER = 4; 7444 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile file_; 7445 /** 7446 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7447 */ 7448 public boolean hasFile() { 7449 return ((bitField0_ & 0x00000008) == 0x00000008); 7450 } 7451 /** 7452 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7453 */ 7454 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile() { 7455 return file_; 7456 } 7457 /** 7458 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7459 */ 7460 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder() { 7461 return file_; 7462 } 7463 7464 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 7465 public static final int DIRECTORY_FIELD_NUMBER = 5; 7466 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory directory_; 7467 /** 7468 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7469 */ 7470 public boolean hasDirectory() { 7471 return ((bitField0_ & 0x00000010) == 0x00000010); 7472 } 7473 /** 7474 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7475 */ 7476 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory() { 7477 return directory_; 7478 } 7479 /** 7480 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 7481 */ 7482 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder() { 7483 return directory_; 7484 } 7485 7486 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 7487 public static final int SYMLINK_FIELD_NUMBER = 6; 7488 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink symlink_; 7489 /** 7490 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7491 */ 7492 public boolean hasSymlink() { 7493 return ((bitField0_ & 0x00000020) == 0x00000020); 7494 } 7495 /** 7496 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7497 */ 7498 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink() { 7499 return symlink_; 7500 } 7501 /** 7502 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 7503 */ 7504 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder() { 7505 return symlink_; 7506 } 7507 7508 private void initFields() { 7509 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 7510 id_ = 0L; 7511 name_ = com.google.protobuf.ByteString.EMPTY; 7512 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 7513 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 7514 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 7515 } 7516 private byte memoizedIsInitialized = -1; 7517 public final boolean isInitialized() { 7518 byte isInitialized = memoizedIsInitialized; 7519 if (isInitialized != -1) return isInitialized == 1; 7520 7521 if (!hasType()) { 7522 memoizedIsInitialized = 0; 7523 return false; 7524 } 7525 if (!hasId()) { 7526 memoizedIsInitialized = 0; 7527 return false; 7528 } 7529 if (hasFile()) { 7530 if (!getFile().isInitialized()) { 7531 memoizedIsInitialized = 0; 7532 return false; 7533 } 7534 } 7535 memoizedIsInitialized = 1; 7536 return true; 7537 } 7538 7539 public void writeTo(com.google.protobuf.CodedOutputStream output) 7540 throws java.io.IOException { 7541 getSerializedSize(); 7542 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7543 output.writeEnum(1, type_.getNumber()); 7544 } 7545 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7546 output.writeUInt64(2, id_); 7547 } 7548 if (((bitField0_ & 0x00000004) == 0x00000004)) { 7549 output.writeBytes(3, name_); 7550 } 7551 if (((bitField0_ & 0x00000008) == 0x00000008)) { 7552 output.writeMessage(4, file_); 7553 } 7554 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7555 output.writeMessage(5, directory_); 7556 } 7557 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7558 output.writeMessage(6, symlink_); 7559 } 7560 getUnknownFields().writeTo(output); 7561 } 7562 7563 private int memoizedSerializedSize = -1; 7564 public int getSerializedSize() { 7565 int size = memoizedSerializedSize; 7566 if (size != -1) return size; 7567 7568 size = 0; 7569 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7570 size += com.google.protobuf.CodedOutputStream 7571 .computeEnumSize(1, type_.getNumber()); 7572 } 7573 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7574 size += com.google.protobuf.CodedOutputStream 7575 .computeUInt64Size(2, id_); 7576 } 7577 if (((bitField0_ & 0x00000004) == 0x00000004)) { 7578 size += com.google.protobuf.CodedOutputStream 7579 .computeBytesSize(3, name_); 7580 } 7581 if (((bitField0_ & 0x00000008) == 0x00000008)) { 7582 size += com.google.protobuf.CodedOutputStream 7583 .computeMessageSize(4, file_); 7584 } 7585 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7586 size += com.google.protobuf.CodedOutputStream 7587 .computeMessageSize(5, directory_); 7588 } 7589 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7590 size += com.google.protobuf.CodedOutputStream 7591 .computeMessageSize(6, symlink_); 7592 } 7593 size += getUnknownFields().getSerializedSize(); 7594 memoizedSerializedSize = size; 7595 return size; 7596 } 7597 7598 private static final long serialVersionUID = 0L; 7599 @java.lang.Override 7600 protected java.lang.Object writeReplace() 7601 throws java.io.ObjectStreamException { 7602 return super.writeReplace(); 7603 } 7604 7605 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7606 com.google.protobuf.ByteString data) 7607 throws com.google.protobuf.InvalidProtocolBufferException { 7608 return PARSER.parseFrom(data); 7609 } 7610 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7611 com.google.protobuf.ByteString data, 7612 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7613 throws com.google.protobuf.InvalidProtocolBufferException { 7614 return PARSER.parseFrom(data, extensionRegistry); 7615 } 7616 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom(byte[] data) 7617 throws com.google.protobuf.InvalidProtocolBufferException { 7618 return PARSER.parseFrom(data); 7619 } 7620 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7621 byte[] data, 7622 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7623 throws com.google.protobuf.InvalidProtocolBufferException { 7624 return PARSER.parseFrom(data, extensionRegistry); 7625 } 7626 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom(java.io.InputStream input) 7627 throws java.io.IOException { 7628 return PARSER.parseFrom(input); 7629 } 7630 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7631 java.io.InputStream input, 7632 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7633 throws java.io.IOException { 7634 return PARSER.parseFrom(input, extensionRegistry); 7635 } 7636 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseDelimitedFrom(java.io.InputStream input) 7637 throws java.io.IOException { 7638 return PARSER.parseDelimitedFrom(input); 7639 } 7640 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseDelimitedFrom( 7641 java.io.InputStream input, 7642 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7643 throws java.io.IOException { 7644 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7645 } 7646 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7647 com.google.protobuf.CodedInputStream input) 7648 throws java.io.IOException { 7649 return PARSER.parseFrom(input); 7650 } 7651 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 7652 com.google.protobuf.CodedInputStream input, 7653 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7654 throws java.io.IOException { 7655 return PARSER.parseFrom(input, extensionRegistry); 7656 } 7657 7658 public static Builder newBuilder() { return Builder.create(); } 7659 public Builder newBuilderForType() { return newBuilder(); } 7660 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode prototype) { 7661 return newBuilder().mergeFrom(prototype); 7662 } 7663 public Builder toBuilder() { return newBuilder(this); } 7664 7665 @java.lang.Override 7666 protected Builder newBuilderForType( 7667 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7668 Builder builder = new Builder(parent); 7669 return builder; 7670 } 7671 /** 7672 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INode} 7673 */ 7674 public static final class Builder extends 7675 com.google.protobuf.GeneratedMessage.Builder<Builder> 7676 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder { 7677 public static final com.google.protobuf.Descriptors.Descriptor 7678 getDescriptor() { 7679 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 7680 } 7681 7682 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7683 internalGetFieldAccessorTable() { 7684 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable 7685 .ensureFieldAccessorsInitialized( 7686 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder.class); 7687 } 7688 7689 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.newBuilder() 7690 private Builder() { 7691 maybeForceBuilderInitialization(); 7692 } 7693 7694 private Builder( 7695 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7696 super(parent); 7697 maybeForceBuilderInitialization(); 7698 } 7699 private void maybeForceBuilderInitialization() { 7700 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7701 getFileFieldBuilder(); 7702 getDirectoryFieldBuilder(); 7703 getSymlinkFieldBuilder(); 7704 } 7705 } 7706 private static Builder create() { 7707 return new Builder(); 7708 } 7709 7710 public Builder clear() { 7711 super.clear(); 7712 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 7713 bitField0_ = (bitField0_ & ~0x00000001); 7714 id_ = 0L; 7715 bitField0_ = (bitField0_ & ~0x00000002); 7716 name_ = com.google.protobuf.ByteString.EMPTY; 7717 bitField0_ = (bitField0_ & ~0x00000004); 7718 if (fileBuilder_ == null) { 7719 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 7720 } else { 7721 fileBuilder_.clear(); 7722 } 7723 bitField0_ = (bitField0_ & ~0x00000008); 7724 if (directoryBuilder_ == null) { 7725 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 7726 } else { 7727 directoryBuilder_.clear(); 7728 } 7729 bitField0_ = (bitField0_ & ~0x00000010); 7730 if (symlinkBuilder_ == null) { 7731 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 7732 } else { 7733 symlinkBuilder_.clear(); 7734 } 7735 bitField0_ = (bitField0_ & ~0x00000020); 7736 return this; 7737 } 7738 7739 public Builder clone() { 7740 return create().mergeFrom(buildPartial()); 7741 } 7742 7743 public com.google.protobuf.Descriptors.Descriptor 7744 getDescriptorForType() { 7745 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 7746 } 7747 7748 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getDefaultInstanceForType() { 7749 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 7750 } 7751 7752 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode build() { 7753 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode result = buildPartial(); 7754 if (!result.isInitialized()) { 7755 throw newUninitializedMessageException(result); 7756 } 7757 return result; 7758 } 7759 7760 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode buildPartial() { 7761 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode(this); 7762 int from_bitField0_ = bitField0_; 7763 int to_bitField0_ = 0; 7764 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 7765 to_bitField0_ |= 0x00000001; 7766 } 7767 result.type_ = type_; 7768 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 7769 to_bitField0_ |= 0x00000002; 7770 } 7771 result.id_ = id_; 7772 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 7773 to_bitField0_ |= 0x00000004; 7774 } 7775 result.name_ = name_; 7776 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 7777 to_bitField0_ |= 0x00000008; 7778 } 7779 if (fileBuilder_ == null) { 7780 result.file_ = file_; 7781 } else { 7782 result.file_ = fileBuilder_.build(); 7783 } 7784 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 7785 to_bitField0_ |= 0x00000010; 7786 } 7787 if (directoryBuilder_ == null) { 7788 result.directory_ = directory_; 7789 } else { 7790 result.directory_ = directoryBuilder_.build(); 7791 } 7792 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 7793 to_bitField0_ |= 0x00000020; 7794 } 7795 if (symlinkBuilder_ == null) { 7796 result.symlink_ = symlink_; 7797 } else { 7798 result.symlink_ = symlinkBuilder_.build(); 7799 } 7800 result.bitField0_ = to_bitField0_; 7801 onBuilt(); 7802 return result; 7803 } 7804 7805 public Builder mergeFrom(com.google.protobuf.Message other) { 7806 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode) { 7807 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode)other); 7808 } else { 7809 super.mergeFrom(other); 7810 return this; 7811 } 7812 } 7813 7814 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode other) { 7815 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance()) return this; 7816 if (other.hasType()) { 7817 setType(other.getType()); 7818 } 7819 if (other.hasId()) { 7820 setId(other.getId()); 7821 } 7822 if (other.hasName()) { 7823 setName(other.getName()); 7824 } 7825 if (other.hasFile()) { 7826 mergeFile(other.getFile()); 7827 } 7828 if (other.hasDirectory()) { 7829 mergeDirectory(other.getDirectory()); 7830 } 7831 if (other.hasSymlink()) { 7832 mergeSymlink(other.getSymlink()); 7833 } 7834 this.mergeUnknownFields(other.getUnknownFields()); 7835 return this; 7836 } 7837 7838 public final boolean isInitialized() { 7839 if (!hasType()) { 7840 7841 return false; 7842 } 7843 if (!hasId()) { 7844 7845 return false; 7846 } 7847 if (hasFile()) { 7848 if (!getFile().isInitialized()) { 7849 7850 return false; 7851 } 7852 } 7853 return true; 7854 } 7855 7856 public Builder mergeFrom( 7857 com.google.protobuf.CodedInputStream input, 7858 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7859 throws java.io.IOException { 7860 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parsedMessage = null; 7861 try { 7862 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7863 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7864 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode) e.getUnfinishedMessage(); 7865 throw e; 7866 } finally { 7867 if (parsedMessage != null) { 7868 mergeFrom(parsedMessage); 7869 } 7870 } 7871 return this; 7872 } 7873 private int bitField0_; 7874 7875 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 7876 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 7877 /** 7878 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7879 */ 7880 public boolean hasType() { 7881 return ((bitField0_ & 0x00000001) == 0x00000001); 7882 } 7883 /** 7884 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7885 */ 7886 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType() { 7887 return type_; 7888 } 7889 /** 7890 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7891 */ 7892 public Builder setType(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type value) { 7893 if (value == null) { 7894 throw new NullPointerException(); 7895 } 7896 bitField0_ |= 0x00000001; 7897 type_ = value; 7898 onChanged(); 7899 return this; 7900 } 7901 /** 7902 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 7903 */ 7904 public Builder clearType() { 7905 bitField0_ = (bitField0_ & ~0x00000001); 7906 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 7907 onChanged(); 7908 return this; 7909 } 7910 7911 // required uint64 id = 2; 7912 private long id_ ; 7913 /** 7914 * <code>required uint64 id = 2;</code> 7915 */ 7916 public boolean hasId() { 7917 return ((bitField0_ & 0x00000002) == 0x00000002); 7918 } 7919 /** 7920 * <code>required uint64 id = 2;</code> 7921 */ 7922 public long getId() { 7923 return id_; 7924 } 7925 /** 7926 * <code>required uint64 id = 2;</code> 7927 */ 7928 public Builder setId(long value) { 7929 bitField0_ |= 0x00000002; 7930 id_ = value; 7931 onChanged(); 7932 return this; 7933 } 7934 /** 7935 * <code>required uint64 id = 2;</code> 7936 */ 7937 public Builder clearId() { 7938 bitField0_ = (bitField0_ & ~0x00000002); 7939 id_ = 0L; 7940 onChanged(); 7941 return this; 7942 } 7943 7944 // optional bytes name = 3; 7945 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 7946 /** 7947 * <code>optional bytes name = 3;</code> 7948 */ 7949 public boolean hasName() { 7950 return ((bitField0_ & 0x00000004) == 0x00000004); 7951 } 7952 /** 7953 * <code>optional bytes name = 3;</code> 7954 */ 7955 public com.google.protobuf.ByteString getName() { 7956 return name_; 7957 } 7958 /** 7959 * <code>optional bytes name = 3;</code> 7960 */ 7961 public Builder setName(com.google.protobuf.ByteString value) { 7962 if (value == null) { 7963 throw new NullPointerException(); 7964 } 7965 bitField0_ |= 0x00000004; 7966 name_ = value; 7967 onChanged(); 7968 return this; 7969 } 7970 /** 7971 * <code>optional bytes name = 3;</code> 7972 */ 7973 public Builder clearName() { 7974 bitField0_ = (bitField0_ & ~0x00000004); 7975 name_ = getDefaultInstance().getName(); 7976 onChanged(); 7977 return this; 7978 } 7979 7980 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 7981 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 7982 private com.google.protobuf.SingleFieldBuilder< 7983 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> fileBuilder_; 7984 /** 7985 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7986 */ 7987 public boolean hasFile() { 7988 return ((bitField0_ & 0x00000008) == 0x00000008); 7989 } 7990 /** 7991 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 7992 */ 7993 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile() { 7994 if (fileBuilder_ == null) { 7995 return file_; 7996 } else { 7997 return fileBuilder_.getMessage(); 7998 } 7999 } 8000 /** 8001 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8002 */ 8003 public Builder setFile(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 8004 if (fileBuilder_ == null) { 8005 if (value == null) { 8006 throw new NullPointerException(); 8007 } 8008 file_ = value; 8009 onChanged(); 8010 } else { 8011 fileBuilder_.setMessage(value); 8012 } 8013 bitField0_ |= 0x00000008; 8014 return this; 8015 } 8016 /** 8017 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8018 */ 8019 public Builder setFile( 8020 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder builderForValue) { 8021 if (fileBuilder_ == null) { 8022 file_ = builderForValue.build(); 8023 onChanged(); 8024 } else { 8025 fileBuilder_.setMessage(builderForValue.build()); 8026 } 8027 bitField0_ |= 0x00000008; 8028 return this; 8029 } 8030 /** 8031 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8032 */ 8033 public Builder mergeFile(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 8034 if (fileBuilder_ == null) { 8035 if (((bitField0_ & 0x00000008) == 0x00000008) && 8036 file_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) { 8037 file_ = 8038 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder(file_).mergeFrom(value).buildPartial(); 8039 } else { 8040 file_ = value; 8041 } 8042 onChanged(); 8043 } else { 8044 fileBuilder_.mergeFrom(value); 8045 } 8046 bitField0_ |= 0x00000008; 8047 return this; 8048 } 8049 /** 8050 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8051 */ 8052 public Builder clearFile() { 8053 if (fileBuilder_ == null) { 8054 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 8055 onChanged(); 8056 } else { 8057 fileBuilder_.clear(); 8058 } 8059 bitField0_ = (bitField0_ & ~0x00000008); 8060 return this; 8061 } 8062 /** 8063 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8064 */ 8065 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder getFileBuilder() { 8066 bitField0_ |= 0x00000008; 8067 onChanged(); 8068 return getFileFieldBuilder().getBuilder(); 8069 } 8070 /** 8071 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8072 */ 8073 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder() { 8074 if (fileBuilder_ != null) { 8075 return fileBuilder_.getMessageOrBuilder(); 8076 } else { 8077 return file_; 8078 } 8079 } 8080 /** 8081 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8082 */ 8083 private com.google.protobuf.SingleFieldBuilder< 8084 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> 8085 getFileFieldBuilder() { 8086 if (fileBuilder_ == null) { 8087 fileBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8088 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder>( 8089 file_, 8090 getParentForChildren(), 8091 isClean()); 8092 file_ = null; 8093 } 8094 return fileBuilder_; 8095 } 8096 8097 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 8098 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 8099 private com.google.protobuf.SingleFieldBuilder< 8100 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> directoryBuilder_; 8101 /** 8102 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8103 */ 8104 public boolean hasDirectory() { 8105 return ((bitField0_ & 0x00000010) == 0x00000010); 8106 } 8107 /** 8108 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8109 */ 8110 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory() { 8111 if (directoryBuilder_ == null) { 8112 return directory_; 8113 } else { 8114 return directoryBuilder_.getMessage(); 8115 } 8116 } 8117 /** 8118 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8119 */ 8120 public Builder setDirectory(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 8121 if (directoryBuilder_ == null) { 8122 if (value == null) { 8123 throw new NullPointerException(); 8124 } 8125 directory_ = value; 8126 onChanged(); 8127 } else { 8128 directoryBuilder_.setMessage(value); 8129 } 8130 bitField0_ |= 0x00000010; 8131 return this; 8132 } 8133 /** 8134 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8135 */ 8136 public Builder setDirectory( 8137 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder builderForValue) { 8138 if (directoryBuilder_ == null) { 8139 directory_ = builderForValue.build(); 8140 onChanged(); 8141 } else { 8142 directoryBuilder_.setMessage(builderForValue.build()); 8143 } 8144 bitField0_ |= 0x00000010; 8145 return this; 8146 } 8147 /** 8148 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8149 */ 8150 public Builder mergeDirectory(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 8151 if (directoryBuilder_ == null) { 8152 if (((bitField0_ & 0x00000010) == 0x00000010) && 8153 directory_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) { 8154 directory_ = 8155 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder(directory_).mergeFrom(value).buildPartial(); 8156 } else { 8157 directory_ = value; 8158 } 8159 onChanged(); 8160 } else { 8161 directoryBuilder_.mergeFrom(value); 8162 } 8163 bitField0_ |= 0x00000010; 8164 return this; 8165 } 8166 /** 8167 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8168 */ 8169 public Builder clearDirectory() { 8170 if (directoryBuilder_ == null) { 8171 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 8172 onChanged(); 8173 } else { 8174 directoryBuilder_.clear(); 8175 } 8176 bitField0_ = (bitField0_ & ~0x00000010); 8177 return this; 8178 } 8179 /** 8180 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8181 */ 8182 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder getDirectoryBuilder() { 8183 bitField0_ |= 0x00000010; 8184 onChanged(); 8185 return getDirectoryFieldBuilder().getBuilder(); 8186 } 8187 /** 8188 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8189 */ 8190 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder() { 8191 if (directoryBuilder_ != null) { 8192 return directoryBuilder_.getMessageOrBuilder(); 8193 } else { 8194 return directory_; 8195 } 8196 } 8197 /** 8198 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8199 */ 8200 private com.google.protobuf.SingleFieldBuilder< 8201 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> 8202 getDirectoryFieldBuilder() { 8203 if (directoryBuilder_ == null) { 8204 directoryBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8205 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder>( 8206 directory_, 8207 getParentForChildren(), 8208 isClean()); 8209 directory_ = null; 8210 } 8211 return directoryBuilder_; 8212 } 8213 8214 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 8215 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 8216 private com.google.protobuf.SingleFieldBuilder< 8217 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder> symlinkBuilder_; 8218 /** 8219 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8220 */ 8221 public boolean hasSymlink() { 8222 return ((bitField0_ & 0x00000020) == 0x00000020); 8223 } 8224 /** 8225 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8226 */ 8227 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink() { 8228 if (symlinkBuilder_ == null) { 8229 return symlink_; 8230 } else { 8231 return symlinkBuilder_.getMessage(); 8232 } 8233 } 8234 /** 8235 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8236 */ 8237 public Builder setSymlink(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink value) { 8238 if (symlinkBuilder_ == null) { 8239 if (value == null) { 8240 throw new NullPointerException(); 8241 } 8242 symlink_ = value; 8243 onChanged(); 8244 } else { 8245 symlinkBuilder_.setMessage(value); 8246 } 8247 bitField0_ |= 0x00000020; 8248 return this; 8249 } 8250 /** 8251 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8252 */ 8253 public Builder setSymlink( 8254 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder builderForValue) { 8255 if (symlinkBuilder_ == null) { 8256 symlink_ = builderForValue.build(); 8257 onChanged(); 8258 } else { 8259 symlinkBuilder_.setMessage(builderForValue.build()); 8260 } 8261 bitField0_ |= 0x00000020; 8262 return this; 8263 } 8264 /** 8265 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8266 */ 8267 public Builder mergeSymlink(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink value) { 8268 if (symlinkBuilder_ == null) { 8269 if (((bitField0_ & 0x00000020) == 0x00000020) && 8270 symlink_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance()) { 8271 symlink_ = 8272 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.newBuilder(symlink_).mergeFrom(value).buildPartial(); 8273 } else { 8274 symlink_ = value; 8275 } 8276 onChanged(); 8277 } else { 8278 symlinkBuilder_.mergeFrom(value); 8279 } 8280 bitField0_ |= 0x00000020; 8281 return this; 8282 } 8283 /** 8284 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8285 */ 8286 public Builder clearSymlink() { 8287 if (symlinkBuilder_ == null) { 8288 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 8289 onChanged(); 8290 } else { 8291 symlinkBuilder_.clear(); 8292 } 8293 bitField0_ = (bitField0_ & ~0x00000020); 8294 return this; 8295 } 8296 /** 8297 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8298 */ 8299 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder getSymlinkBuilder() { 8300 bitField0_ |= 0x00000020; 8301 onChanged(); 8302 return getSymlinkFieldBuilder().getBuilder(); 8303 } 8304 /** 8305 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8306 */ 8307 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder() { 8308 if (symlinkBuilder_ != null) { 8309 return symlinkBuilder_.getMessageOrBuilder(); 8310 } else { 8311 return symlink_; 8312 } 8313 } 8314 /** 8315 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8316 */ 8317 private com.google.protobuf.SingleFieldBuilder< 8318 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder> 8319 getSymlinkFieldBuilder() { 8320 if (symlinkBuilder_ == null) { 8321 symlinkBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8322 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder>( 8323 symlink_, 8324 getParentForChildren(), 8325 isClean()); 8326 symlink_ = null; 8327 } 8328 return symlinkBuilder_; 8329 } 8330 8331 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INode) 8332 } 8333 8334 static { 8335 defaultInstance = new INode(true); 8336 defaultInstance.initFields(); 8337 } 8338 8339 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INode) 8340 } 8341 8342 private int bitField0_; 8343 // optional uint64 lastInodeId = 1; 8344 public static final int LASTINODEID_FIELD_NUMBER = 1; 8345 private long lastInodeId_; 8346 /** 8347 * <code>optional uint64 lastInodeId = 1;</code> 8348 */ 8349 public boolean hasLastInodeId() { 8350 return ((bitField0_ & 0x00000001) == 0x00000001); 8351 } 8352 /** 8353 * <code>optional uint64 lastInodeId = 1;</code> 8354 */ 8355 public long getLastInodeId() { 8356 return lastInodeId_; 8357 } 8358 8359 // optional uint64 numInodes = 2; 8360 public static final int NUMINODES_FIELD_NUMBER = 2; 8361 private long numInodes_; 8362 /** 8363 * <code>optional uint64 numInodes = 2;</code> 8364 * 8365 * <pre> 8366 * repeated INodes.. 8367 * </pre> 8368 */ 8369 public boolean hasNumInodes() { 8370 return ((bitField0_ & 0x00000002) == 0x00000002); 8371 } 8372 /** 8373 * <code>optional uint64 numInodes = 2;</code> 8374 * 8375 * <pre> 8376 * repeated INodes.. 8377 * </pre> 8378 */ 8379 public long getNumInodes() { 8380 return numInodes_; 8381 } 8382 8383 private void initFields() { 8384 lastInodeId_ = 0L; 8385 numInodes_ = 0L; 8386 } 8387 private byte memoizedIsInitialized = -1; 8388 public final boolean isInitialized() { 8389 byte isInitialized = memoizedIsInitialized; 8390 if (isInitialized != -1) return isInitialized == 1; 8391 8392 memoizedIsInitialized = 1; 8393 return true; 8394 } 8395 8396 public void writeTo(com.google.protobuf.CodedOutputStream output) 8397 throws java.io.IOException { 8398 getSerializedSize(); 8399 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8400 output.writeUInt64(1, lastInodeId_); 8401 } 8402 if (((bitField0_ & 0x00000002) == 0x00000002)) { 8403 output.writeUInt64(2, numInodes_); 8404 } 8405 getUnknownFields().writeTo(output); 8406 } 8407 8408 private int memoizedSerializedSize = -1; 8409 public int getSerializedSize() { 8410 int size = memoizedSerializedSize; 8411 if (size != -1) return size; 8412 8413 size = 0; 8414 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8415 size += com.google.protobuf.CodedOutputStream 8416 .computeUInt64Size(1, lastInodeId_); 8417 } 8418 if (((bitField0_ & 0x00000002) == 0x00000002)) { 8419 size += com.google.protobuf.CodedOutputStream 8420 .computeUInt64Size(2, numInodes_); 8421 } 8422 size += getUnknownFields().getSerializedSize(); 8423 memoizedSerializedSize = size; 8424 return size; 8425 } 8426 8427 private static final long serialVersionUID = 0L; 8428 @java.lang.Override 8429 protected java.lang.Object writeReplace() 8430 throws java.io.ObjectStreamException { 8431 return super.writeReplace(); 8432 } 8433 8434 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8435 com.google.protobuf.ByteString data) 8436 throws com.google.protobuf.InvalidProtocolBufferException { 8437 return PARSER.parseFrom(data); 8438 } 8439 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8440 com.google.protobuf.ByteString data, 8441 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8442 throws com.google.protobuf.InvalidProtocolBufferException { 8443 return PARSER.parseFrom(data, extensionRegistry); 8444 } 8445 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom(byte[] data) 8446 throws com.google.protobuf.InvalidProtocolBufferException { 8447 return PARSER.parseFrom(data); 8448 } 8449 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8450 byte[] data, 8451 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8452 throws com.google.protobuf.InvalidProtocolBufferException { 8453 return PARSER.parseFrom(data, extensionRegistry); 8454 } 8455 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom(java.io.InputStream input) 8456 throws java.io.IOException { 8457 return PARSER.parseFrom(input); 8458 } 8459 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8460 java.io.InputStream input, 8461 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8462 throws java.io.IOException { 8463 return PARSER.parseFrom(input, extensionRegistry); 8464 } 8465 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseDelimitedFrom(java.io.InputStream input) 8466 throws java.io.IOException { 8467 return PARSER.parseDelimitedFrom(input); 8468 } 8469 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseDelimitedFrom( 8470 java.io.InputStream input, 8471 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8472 throws java.io.IOException { 8473 return PARSER.parseDelimitedFrom(input, extensionRegistry); 8474 } 8475 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8476 com.google.protobuf.CodedInputStream input) 8477 throws java.io.IOException { 8478 return PARSER.parseFrom(input); 8479 } 8480 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 8481 com.google.protobuf.CodedInputStream input, 8482 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8483 throws java.io.IOException { 8484 return PARSER.parseFrom(input, extensionRegistry); 8485 } 8486 8487 public static Builder newBuilder() { return Builder.create(); } 8488 public Builder newBuilderForType() { return newBuilder(); } 8489 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection prototype) { 8490 return newBuilder().mergeFrom(prototype); 8491 } 8492 public Builder toBuilder() { return newBuilder(this); } 8493 8494 @java.lang.Override 8495 protected Builder newBuilderForType( 8496 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8497 Builder builder = new Builder(parent); 8498 return builder; 8499 } 8500 /** 8501 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection} 8502 * 8503 * <pre> 8504 ** 8505 * Permission is serialized as a 64-bit long. [0:24):[25:48):[48:64) (in Big Endian). 8506 * The first and the second parts are the string ids of the user and 8507 * group name, and the last 16 bits are the permission bits. 8508 * 8509 * Name: INODE 8510 * </pre> 8511 */ 8512 public static final class Builder extends 8513 com.google.protobuf.GeneratedMessage.Builder<Builder> 8514 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSectionOrBuilder { 8515 public static final com.google.protobuf.Descriptors.Descriptor 8516 getDescriptor() { 8517 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 8518 } 8519 8520 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8521 internalGetFieldAccessorTable() { 8522 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable 8523 .ensureFieldAccessorsInitialized( 8524 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.Builder.class); 8525 } 8526 8527 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.newBuilder() 8528 private Builder() { 8529 maybeForceBuilderInitialization(); 8530 } 8531 8532 private Builder( 8533 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8534 super(parent); 8535 maybeForceBuilderInitialization(); 8536 } 8537 private void maybeForceBuilderInitialization() { 8538 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8539 } 8540 } 8541 private static Builder create() { 8542 return new Builder(); 8543 } 8544 8545 public Builder clear() { 8546 super.clear(); 8547 lastInodeId_ = 0L; 8548 bitField0_ = (bitField0_ & ~0x00000001); 8549 numInodes_ = 0L; 8550 bitField0_ = (bitField0_ & ~0x00000002); 8551 return this; 8552 } 8553 8554 public Builder clone() { 8555 return create().mergeFrom(buildPartial()); 8556 } 8557 8558 public com.google.protobuf.Descriptors.Descriptor 8559 getDescriptorForType() { 8560 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 8561 } 8562 8563 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection getDefaultInstanceForType() { 8564 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.getDefaultInstance(); 8565 } 8566 8567 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection build() { 8568 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection result = buildPartial(); 8569 if (!result.isInitialized()) { 8570 throw newUninitializedMessageException(result); 8571 } 8572 return result; 8573 } 8574 8575 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection buildPartial() { 8576 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection(this); 8577 int from_bitField0_ = bitField0_; 8578 int to_bitField0_ = 0; 8579 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 8580 to_bitField0_ |= 0x00000001; 8581 } 8582 result.lastInodeId_ = lastInodeId_; 8583 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 8584 to_bitField0_ |= 0x00000002; 8585 } 8586 result.numInodes_ = numInodes_; 8587 result.bitField0_ = to_bitField0_; 8588 onBuilt(); 8589 return result; 8590 } 8591 8592 public Builder mergeFrom(com.google.protobuf.Message other) { 8593 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection) { 8594 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection)other); 8595 } else { 8596 super.mergeFrom(other); 8597 return this; 8598 } 8599 } 8600 8601 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection other) { 8602 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.getDefaultInstance()) return this; 8603 if (other.hasLastInodeId()) { 8604 setLastInodeId(other.getLastInodeId()); 8605 } 8606 if (other.hasNumInodes()) { 8607 setNumInodes(other.getNumInodes()); 8608 } 8609 this.mergeUnknownFields(other.getUnknownFields()); 8610 return this; 8611 } 8612 8613 public final boolean isInitialized() { 8614 return true; 8615 } 8616 8617 public Builder mergeFrom( 8618 com.google.protobuf.CodedInputStream input, 8619 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8620 throws java.io.IOException { 8621 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parsedMessage = null; 8622 try { 8623 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 8624 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8625 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection) e.getUnfinishedMessage(); 8626 throw e; 8627 } finally { 8628 if (parsedMessage != null) { 8629 mergeFrom(parsedMessage); 8630 } 8631 } 8632 return this; 8633 } 8634 private int bitField0_; 8635 8636 // optional uint64 lastInodeId = 1; 8637 private long lastInodeId_ ; 8638 /** 8639 * <code>optional uint64 lastInodeId = 1;</code> 8640 */ 8641 public boolean hasLastInodeId() { 8642 return ((bitField0_ & 0x00000001) == 0x00000001); 8643 } 8644 /** 8645 * <code>optional uint64 lastInodeId = 1;</code> 8646 */ 8647 public long getLastInodeId() { 8648 return lastInodeId_; 8649 } 8650 /** 8651 * <code>optional uint64 lastInodeId = 1;</code> 8652 */ 8653 public Builder setLastInodeId(long value) { 8654 bitField0_ |= 0x00000001; 8655 lastInodeId_ = value; 8656 onChanged(); 8657 return this; 8658 } 8659 /** 8660 * <code>optional uint64 lastInodeId = 1;</code> 8661 */ 8662 public Builder clearLastInodeId() { 8663 bitField0_ = (bitField0_ & ~0x00000001); 8664 lastInodeId_ = 0L; 8665 onChanged(); 8666 return this; 8667 } 8668 8669 // optional uint64 numInodes = 2; 8670 private long numInodes_ ; 8671 /** 8672 * <code>optional uint64 numInodes = 2;</code> 8673 * 8674 * <pre> 8675 * repeated INodes.. 8676 * </pre> 8677 */ 8678 public boolean hasNumInodes() { 8679 return ((bitField0_ & 0x00000002) == 0x00000002); 8680 } 8681 /** 8682 * <code>optional uint64 numInodes = 2;</code> 8683 * 8684 * <pre> 8685 * repeated INodes.. 8686 * </pre> 8687 */ 8688 public long getNumInodes() { 8689 return numInodes_; 8690 } 8691 /** 8692 * <code>optional uint64 numInodes = 2;</code> 8693 * 8694 * <pre> 8695 * repeated INodes.. 8696 * </pre> 8697 */ 8698 public Builder setNumInodes(long value) { 8699 bitField0_ |= 0x00000002; 8700 numInodes_ = value; 8701 onChanged(); 8702 return this; 8703 } 8704 /** 8705 * <code>optional uint64 numInodes = 2;</code> 8706 * 8707 * <pre> 8708 * repeated INodes.. 8709 * </pre> 8710 */ 8711 public Builder clearNumInodes() { 8712 bitField0_ = (bitField0_ & ~0x00000002); 8713 numInodes_ = 0L; 8714 onChanged(); 8715 return this; 8716 } 8717 8718 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection) 8719 } 8720 8721 static { 8722 defaultInstance = new INodeSection(true); 8723 defaultInstance.initFields(); 8724 } 8725 8726 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection) 8727 } 8728 8729 public interface FilesUnderConstructionSectionOrBuilder 8730 extends com.google.protobuf.MessageOrBuilder { 8731 } 8732 /** 8733 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection} 8734 * 8735 * <pre> 8736 ** 8737 * This section records information about under-construction files for 8738 * reconstructing the lease map. 8739 * NAME: FILES_UNDERCONSTRUCTION 8740 * </pre> 8741 */ 8742 public static final class FilesUnderConstructionSection extends 8743 com.google.protobuf.GeneratedMessage 8744 implements FilesUnderConstructionSectionOrBuilder { 8745 // Use FilesUnderConstructionSection.newBuilder() to construct. 8746 private FilesUnderConstructionSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8747 super(builder); 8748 this.unknownFields = builder.getUnknownFields(); 8749 } 8750 private FilesUnderConstructionSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8751 8752 private static final FilesUnderConstructionSection defaultInstance; 8753 public static FilesUnderConstructionSection getDefaultInstance() { 8754 return defaultInstance; 8755 } 8756 8757 public FilesUnderConstructionSection getDefaultInstanceForType() { 8758 return defaultInstance; 8759 } 8760 8761 private final com.google.protobuf.UnknownFieldSet unknownFields; 8762 @java.lang.Override 8763 public final com.google.protobuf.UnknownFieldSet 8764 getUnknownFields() { 8765 return this.unknownFields; 8766 } 8767 private FilesUnderConstructionSection( 8768 com.google.protobuf.CodedInputStream input, 8769 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8770 throws com.google.protobuf.InvalidProtocolBufferException { 8771 initFields(); 8772 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8773 com.google.protobuf.UnknownFieldSet.newBuilder(); 8774 try { 8775 boolean done = false; 8776 while (!done) { 8777 int tag = input.readTag(); 8778 switch (tag) { 8779 case 0: 8780 done = true; 8781 break; 8782 default: { 8783 if (!parseUnknownField(input, unknownFields, 8784 extensionRegistry, tag)) { 8785 done = true; 8786 } 8787 break; 8788 } 8789 } 8790 } 8791 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8792 throw e.setUnfinishedMessage(this); 8793 } catch (java.io.IOException e) { 8794 throw new com.google.protobuf.InvalidProtocolBufferException( 8795 e.getMessage()).setUnfinishedMessage(this); 8796 } finally { 8797 this.unknownFields = unknownFields.build(); 8798 makeExtensionsImmutable(); 8799 } 8800 } 8801 public static final com.google.protobuf.Descriptors.Descriptor 8802 getDescriptor() { 8803 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 8804 } 8805 8806 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8807 internalGetFieldAccessorTable() { 8808 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable 8809 .ensureFieldAccessorsInitialized( 8810 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.Builder.class); 8811 } 8812 8813 public static com.google.protobuf.Parser<FilesUnderConstructionSection> PARSER = 8814 new com.google.protobuf.AbstractParser<FilesUnderConstructionSection>() { 8815 public FilesUnderConstructionSection parsePartialFrom( 8816 com.google.protobuf.CodedInputStream input, 8817 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8818 throws com.google.protobuf.InvalidProtocolBufferException { 8819 return new FilesUnderConstructionSection(input, extensionRegistry); 8820 } 8821 }; 8822 8823 @java.lang.Override 8824 public com.google.protobuf.Parser<FilesUnderConstructionSection> getParserForType() { 8825 return PARSER; 8826 } 8827 8828 public interface FileUnderConstructionEntryOrBuilder 8829 extends com.google.protobuf.MessageOrBuilder { 8830 8831 // optional uint64 inodeId = 1; 8832 /** 8833 * <code>optional uint64 inodeId = 1;</code> 8834 */ 8835 boolean hasInodeId(); 8836 /** 8837 * <code>optional uint64 inodeId = 1;</code> 8838 */ 8839 long getInodeId(); 8840 8841 // optional string fullPath = 2; 8842 /** 8843 * <code>optional string fullPath = 2;</code> 8844 */ 8845 boolean hasFullPath(); 8846 /** 8847 * <code>optional string fullPath = 2;</code> 8848 */ 8849 java.lang.String getFullPath(); 8850 /** 8851 * <code>optional string fullPath = 2;</code> 8852 */ 8853 com.google.protobuf.ByteString 8854 getFullPathBytes(); 8855 } 8856 /** 8857 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry} 8858 */ 8859 public static final class FileUnderConstructionEntry extends 8860 com.google.protobuf.GeneratedMessage 8861 implements FileUnderConstructionEntryOrBuilder { 8862 // Use FileUnderConstructionEntry.newBuilder() to construct. 8863 private FileUnderConstructionEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8864 super(builder); 8865 this.unknownFields = builder.getUnknownFields(); 8866 } 8867 private FileUnderConstructionEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8868 8869 private static final FileUnderConstructionEntry defaultInstance; 8870 public static FileUnderConstructionEntry getDefaultInstance() { 8871 return defaultInstance; 8872 } 8873 8874 public FileUnderConstructionEntry getDefaultInstanceForType() { 8875 return defaultInstance; 8876 } 8877 8878 private final com.google.protobuf.UnknownFieldSet unknownFields; 8879 @java.lang.Override 8880 public final com.google.protobuf.UnknownFieldSet 8881 getUnknownFields() { 8882 return this.unknownFields; 8883 } 8884 private FileUnderConstructionEntry( 8885 com.google.protobuf.CodedInputStream input, 8886 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8887 throws com.google.protobuf.InvalidProtocolBufferException { 8888 initFields(); 8889 int mutable_bitField0_ = 0; 8890 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8891 com.google.protobuf.UnknownFieldSet.newBuilder(); 8892 try { 8893 boolean done = false; 8894 while (!done) { 8895 int tag = input.readTag(); 8896 switch (tag) { 8897 case 0: 8898 done = true; 8899 break; 8900 default: { 8901 if (!parseUnknownField(input, unknownFields, 8902 extensionRegistry, tag)) { 8903 done = true; 8904 } 8905 break; 8906 } 8907 case 8: { 8908 bitField0_ |= 0x00000001; 8909 inodeId_ = input.readUInt64(); 8910 break; 8911 } 8912 case 18: { 8913 bitField0_ |= 0x00000002; 8914 fullPath_ = input.readBytes(); 8915 break; 8916 } 8917 } 8918 } 8919 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8920 throw e.setUnfinishedMessage(this); 8921 } catch (java.io.IOException e) { 8922 throw new com.google.protobuf.InvalidProtocolBufferException( 8923 e.getMessage()).setUnfinishedMessage(this); 8924 } finally { 8925 this.unknownFields = unknownFields.build(); 8926 makeExtensionsImmutable(); 8927 } 8928 } 8929 public static final com.google.protobuf.Descriptors.Descriptor 8930 getDescriptor() { 8931 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 8932 } 8933 8934 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8935 internalGetFieldAccessorTable() { 8936 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable 8937 .ensureFieldAccessorsInitialized( 8938 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.Builder.class); 8939 } 8940 8941 public static com.google.protobuf.Parser<FileUnderConstructionEntry> PARSER = 8942 new com.google.protobuf.AbstractParser<FileUnderConstructionEntry>() { 8943 public FileUnderConstructionEntry parsePartialFrom( 8944 com.google.protobuf.CodedInputStream input, 8945 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8946 throws com.google.protobuf.InvalidProtocolBufferException { 8947 return new FileUnderConstructionEntry(input, extensionRegistry); 8948 } 8949 }; 8950 8951 @java.lang.Override 8952 public com.google.protobuf.Parser<FileUnderConstructionEntry> getParserForType() { 8953 return PARSER; 8954 } 8955 8956 private int bitField0_; 8957 // optional uint64 inodeId = 1; 8958 public static final int INODEID_FIELD_NUMBER = 1; 8959 private long inodeId_; 8960 /** 8961 * <code>optional uint64 inodeId = 1;</code> 8962 */ 8963 public boolean hasInodeId() { 8964 return ((bitField0_ & 0x00000001) == 0x00000001); 8965 } 8966 /** 8967 * <code>optional uint64 inodeId = 1;</code> 8968 */ 8969 public long getInodeId() { 8970 return inodeId_; 8971 } 8972 8973 // optional string fullPath = 2; 8974 public static final int FULLPATH_FIELD_NUMBER = 2; 8975 private java.lang.Object fullPath_; 8976 /** 8977 * <code>optional string fullPath = 2;</code> 8978 */ 8979 public boolean hasFullPath() { 8980 return ((bitField0_ & 0x00000002) == 0x00000002); 8981 } 8982 /** 8983 * <code>optional string fullPath = 2;</code> 8984 */ 8985 public java.lang.String getFullPath() { 8986 java.lang.Object ref = fullPath_; 8987 if (ref instanceof java.lang.String) { 8988 return (java.lang.String) ref; 8989 } else { 8990 com.google.protobuf.ByteString bs = 8991 (com.google.protobuf.ByteString) ref; 8992 java.lang.String s = bs.toStringUtf8(); 8993 if (bs.isValidUtf8()) { 8994 fullPath_ = s; 8995 } 8996 return s; 8997 } 8998 } 8999 /** 9000 * <code>optional string fullPath = 2;</code> 9001 */ 9002 public com.google.protobuf.ByteString 9003 getFullPathBytes() { 9004 java.lang.Object ref = fullPath_; 9005 if (ref instanceof java.lang.String) { 9006 com.google.protobuf.ByteString b = 9007 com.google.protobuf.ByteString.copyFromUtf8( 9008 (java.lang.String) ref); 9009 fullPath_ = b; 9010 return b; 9011 } else { 9012 return (com.google.protobuf.ByteString) ref; 9013 } 9014 } 9015 9016 private void initFields() { 9017 inodeId_ = 0L; 9018 fullPath_ = ""; 9019 } 9020 private byte memoizedIsInitialized = -1; 9021 public final boolean isInitialized() { 9022 byte isInitialized = memoizedIsInitialized; 9023 if (isInitialized != -1) return isInitialized == 1; 9024 9025 memoizedIsInitialized = 1; 9026 return true; 9027 } 9028 9029 public void writeTo(com.google.protobuf.CodedOutputStream output) 9030 throws java.io.IOException { 9031 getSerializedSize(); 9032 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9033 output.writeUInt64(1, inodeId_); 9034 } 9035 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9036 output.writeBytes(2, getFullPathBytes()); 9037 } 9038 getUnknownFields().writeTo(output); 9039 } 9040 9041 private int memoizedSerializedSize = -1; 9042 public int getSerializedSize() { 9043 int size = memoizedSerializedSize; 9044 if (size != -1) return size; 9045 9046 size = 0; 9047 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9048 size += com.google.protobuf.CodedOutputStream 9049 .computeUInt64Size(1, inodeId_); 9050 } 9051 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9052 size += com.google.protobuf.CodedOutputStream 9053 .computeBytesSize(2, getFullPathBytes()); 9054 } 9055 size += getUnknownFields().getSerializedSize(); 9056 memoizedSerializedSize = size; 9057 return size; 9058 } 9059 9060 private static final long serialVersionUID = 0L; 9061 @java.lang.Override 9062 protected java.lang.Object writeReplace() 9063 throws java.io.ObjectStreamException { 9064 return super.writeReplace(); 9065 } 9066 9067 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9068 com.google.protobuf.ByteString data) 9069 throws com.google.protobuf.InvalidProtocolBufferException { 9070 return PARSER.parseFrom(data); 9071 } 9072 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9073 com.google.protobuf.ByteString data, 9074 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9075 throws com.google.protobuf.InvalidProtocolBufferException { 9076 return PARSER.parseFrom(data, extensionRegistry); 9077 } 9078 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom(byte[] data) 9079 throws com.google.protobuf.InvalidProtocolBufferException { 9080 return PARSER.parseFrom(data); 9081 } 9082 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9083 byte[] data, 9084 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9085 throws com.google.protobuf.InvalidProtocolBufferException { 9086 return PARSER.parseFrom(data, extensionRegistry); 9087 } 9088 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom(java.io.InputStream input) 9089 throws java.io.IOException { 9090 return PARSER.parseFrom(input); 9091 } 9092 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9093 java.io.InputStream input, 9094 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9095 throws java.io.IOException { 9096 return PARSER.parseFrom(input, extensionRegistry); 9097 } 9098 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseDelimitedFrom(java.io.InputStream input) 9099 throws java.io.IOException { 9100 return PARSER.parseDelimitedFrom(input); 9101 } 9102 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseDelimitedFrom( 9103 java.io.InputStream input, 9104 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9105 throws java.io.IOException { 9106 return PARSER.parseDelimitedFrom(input, extensionRegistry); 9107 } 9108 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9109 com.google.protobuf.CodedInputStream input) 9110 throws java.io.IOException { 9111 return PARSER.parseFrom(input); 9112 } 9113 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 9114 com.google.protobuf.CodedInputStream input, 9115 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9116 throws java.io.IOException { 9117 return PARSER.parseFrom(input, extensionRegistry); 9118 } 9119 9120 public static Builder newBuilder() { return Builder.create(); } 9121 public Builder newBuilderForType() { return newBuilder(); } 9122 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry prototype) { 9123 return newBuilder().mergeFrom(prototype); 9124 } 9125 public Builder toBuilder() { return newBuilder(this); } 9126 9127 @java.lang.Override 9128 protected Builder newBuilderForType( 9129 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9130 Builder builder = new Builder(parent); 9131 return builder; 9132 } 9133 /** 9134 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry} 9135 */ 9136 public static final class Builder extends 9137 com.google.protobuf.GeneratedMessage.Builder<Builder> 9138 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntryOrBuilder { 9139 public static final com.google.protobuf.Descriptors.Descriptor 9140 getDescriptor() { 9141 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 9142 } 9143 9144 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9145 internalGetFieldAccessorTable() { 9146 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable 9147 .ensureFieldAccessorsInitialized( 9148 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.Builder.class); 9149 } 9150 9151 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.newBuilder() 9152 private Builder() { 9153 maybeForceBuilderInitialization(); 9154 } 9155 9156 private Builder( 9157 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9158 super(parent); 9159 maybeForceBuilderInitialization(); 9160 } 9161 private void maybeForceBuilderInitialization() { 9162 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 9163 } 9164 } 9165 private static Builder create() { 9166 return new Builder(); 9167 } 9168 9169 public Builder clear() { 9170 super.clear(); 9171 inodeId_ = 0L; 9172 bitField0_ = (bitField0_ & ~0x00000001); 9173 fullPath_ = ""; 9174 bitField0_ = (bitField0_ & ~0x00000002); 9175 return this; 9176 } 9177 9178 public Builder clone() { 9179 return create().mergeFrom(buildPartial()); 9180 } 9181 9182 public com.google.protobuf.Descriptors.Descriptor 9183 getDescriptorForType() { 9184 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 9185 } 9186 9187 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry getDefaultInstanceForType() { 9188 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.getDefaultInstance(); 9189 } 9190 9191 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry build() { 9192 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry result = buildPartial(); 9193 if (!result.isInitialized()) { 9194 throw newUninitializedMessageException(result); 9195 } 9196 return result; 9197 } 9198 9199 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry buildPartial() { 9200 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry(this); 9201 int from_bitField0_ = bitField0_; 9202 int to_bitField0_ = 0; 9203 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 9204 to_bitField0_ |= 0x00000001; 9205 } 9206 result.inodeId_ = inodeId_; 9207 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 9208 to_bitField0_ |= 0x00000002; 9209 } 9210 result.fullPath_ = fullPath_; 9211 result.bitField0_ = to_bitField0_; 9212 onBuilt(); 9213 return result; 9214 } 9215 9216 public Builder mergeFrom(com.google.protobuf.Message other) { 9217 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry) { 9218 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry)other); 9219 } else { 9220 super.mergeFrom(other); 9221 return this; 9222 } 9223 } 9224 9225 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry other) { 9226 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.getDefaultInstance()) return this; 9227 if (other.hasInodeId()) { 9228 setInodeId(other.getInodeId()); 9229 } 9230 if (other.hasFullPath()) { 9231 bitField0_ |= 0x00000002; 9232 fullPath_ = other.fullPath_; 9233 onChanged(); 9234 } 9235 this.mergeUnknownFields(other.getUnknownFields()); 9236 return this; 9237 } 9238 9239 public final boolean isInitialized() { 9240 return true; 9241 } 9242 9243 public Builder mergeFrom( 9244 com.google.protobuf.CodedInputStream input, 9245 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9246 throws java.io.IOException { 9247 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parsedMessage = null; 9248 try { 9249 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 9250 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9251 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry) e.getUnfinishedMessage(); 9252 throw e; 9253 } finally { 9254 if (parsedMessage != null) { 9255 mergeFrom(parsedMessage); 9256 } 9257 } 9258 return this; 9259 } 9260 private int bitField0_; 9261 9262 // optional uint64 inodeId = 1; 9263 private long inodeId_ ; 9264 /** 9265 * <code>optional uint64 inodeId = 1;</code> 9266 */ 9267 public boolean hasInodeId() { 9268 return ((bitField0_ & 0x00000001) == 0x00000001); 9269 } 9270 /** 9271 * <code>optional uint64 inodeId = 1;</code> 9272 */ 9273 public long getInodeId() { 9274 return inodeId_; 9275 } 9276 /** 9277 * <code>optional uint64 inodeId = 1;</code> 9278 */ 9279 public Builder setInodeId(long value) { 9280 bitField0_ |= 0x00000001; 9281 inodeId_ = value; 9282 onChanged(); 9283 return this; 9284 } 9285 /** 9286 * <code>optional uint64 inodeId = 1;</code> 9287 */ 9288 public Builder clearInodeId() { 9289 bitField0_ = (bitField0_ & ~0x00000001); 9290 inodeId_ = 0L; 9291 onChanged(); 9292 return this; 9293 } 9294 9295 // optional string fullPath = 2; 9296 private java.lang.Object fullPath_ = ""; 9297 /** 9298 * <code>optional string fullPath = 2;</code> 9299 */ 9300 public boolean hasFullPath() { 9301 return ((bitField0_ & 0x00000002) == 0x00000002); 9302 } 9303 /** 9304 * <code>optional string fullPath = 2;</code> 9305 */ 9306 public java.lang.String getFullPath() { 9307 java.lang.Object ref = fullPath_; 9308 if (!(ref instanceof java.lang.String)) { 9309 java.lang.String s = ((com.google.protobuf.ByteString) ref) 9310 .toStringUtf8(); 9311 fullPath_ = s; 9312 return s; 9313 } else { 9314 return (java.lang.String) ref; 9315 } 9316 } 9317 /** 9318 * <code>optional string fullPath = 2;</code> 9319 */ 9320 public com.google.protobuf.ByteString 9321 getFullPathBytes() { 9322 java.lang.Object ref = fullPath_; 9323 if (ref instanceof String) { 9324 com.google.protobuf.ByteString b = 9325 com.google.protobuf.ByteString.copyFromUtf8( 9326 (java.lang.String) ref); 9327 fullPath_ = b; 9328 return b; 9329 } else { 9330 return (com.google.protobuf.ByteString) ref; 9331 } 9332 } 9333 /** 9334 * <code>optional string fullPath = 2;</code> 9335 */ 9336 public Builder setFullPath( 9337 java.lang.String value) { 9338 if (value == null) { 9339 throw new NullPointerException(); 9340 } 9341 bitField0_ |= 0x00000002; 9342 fullPath_ = value; 9343 onChanged(); 9344 return this; 9345 } 9346 /** 9347 * <code>optional string fullPath = 2;</code> 9348 */ 9349 public Builder clearFullPath() { 9350 bitField0_ = (bitField0_ & ~0x00000002); 9351 fullPath_ = getDefaultInstance().getFullPath(); 9352 onChanged(); 9353 return this; 9354 } 9355 /** 9356 * <code>optional string fullPath = 2;</code> 9357 */ 9358 public Builder setFullPathBytes( 9359 com.google.protobuf.ByteString value) { 9360 if (value == null) { 9361 throw new NullPointerException(); 9362 } 9363 bitField0_ |= 0x00000002; 9364 fullPath_ = value; 9365 onChanged(); 9366 return this; 9367 } 9368 9369 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry) 9370 } 9371 9372 static { 9373 defaultInstance = new FileUnderConstructionEntry(true); 9374 defaultInstance.initFields(); 9375 } 9376 9377 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry) 9378 } 9379 9380 private void initFields() { 9381 } 9382 private byte memoizedIsInitialized = -1; 9383 public final boolean isInitialized() { 9384 byte isInitialized = memoizedIsInitialized; 9385 if (isInitialized != -1) return isInitialized == 1; 9386 9387 memoizedIsInitialized = 1; 9388 return true; 9389 } 9390 9391 public void writeTo(com.google.protobuf.CodedOutputStream output) 9392 throws java.io.IOException { 9393 getSerializedSize(); 9394 getUnknownFields().writeTo(output); 9395 } 9396 9397 private int memoizedSerializedSize = -1; 9398 public int getSerializedSize() { 9399 int size = memoizedSerializedSize; 9400 if (size != -1) return size; 9401 9402 size = 0; 9403 size += getUnknownFields().getSerializedSize(); 9404 memoizedSerializedSize = size; 9405 return size; 9406 } 9407 9408 private static final long serialVersionUID = 0L; 9409 @java.lang.Override 9410 protected java.lang.Object writeReplace() 9411 throws java.io.ObjectStreamException { 9412 return super.writeReplace(); 9413 } 9414 9415 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9416 com.google.protobuf.ByteString data) 9417 throws com.google.protobuf.InvalidProtocolBufferException { 9418 return PARSER.parseFrom(data); 9419 } 9420 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9421 com.google.protobuf.ByteString data, 9422 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9423 throws com.google.protobuf.InvalidProtocolBufferException { 9424 return PARSER.parseFrom(data, extensionRegistry); 9425 } 9426 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom(byte[] data) 9427 throws com.google.protobuf.InvalidProtocolBufferException { 9428 return PARSER.parseFrom(data); 9429 } 9430 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9431 byte[] data, 9432 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9433 throws com.google.protobuf.InvalidProtocolBufferException { 9434 return PARSER.parseFrom(data, extensionRegistry); 9435 } 9436 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom(java.io.InputStream input) 9437 throws java.io.IOException { 9438 return PARSER.parseFrom(input); 9439 } 9440 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9441 java.io.InputStream input, 9442 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9443 throws java.io.IOException { 9444 return PARSER.parseFrom(input, extensionRegistry); 9445 } 9446 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseDelimitedFrom(java.io.InputStream input) 9447 throws java.io.IOException { 9448 return PARSER.parseDelimitedFrom(input); 9449 } 9450 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseDelimitedFrom( 9451 java.io.InputStream input, 9452 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9453 throws java.io.IOException { 9454 return PARSER.parseDelimitedFrom(input, extensionRegistry); 9455 } 9456 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9457 com.google.protobuf.CodedInputStream input) 9458 throws java.io.IOException { 9459 return PARSER.parseFrom(input); 9460 } 9461 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 9462 com.google.protobuf.CodedInputStream input, 9463 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9464 throws java.io.IOException { 9465 return PARSER.parseFrom(input, extensionRegistry); 9466 } 9467 9468 public static Builder newBuilder() { return Builder.create(); } 9469 public Builder newBuilderForType() { return newBuilder(); } 9470 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection prototype) { 9471 return newBuilder().mergeFrom(prototype); 9472 } 9473 public Builder toBuilder() { return newBuilder(this); } 9474 9475 @java.lang.Override 9476 protected Builder newBuilderForType( 9477 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9478 Builder builder = new Builder(parent); 9479 return builder; 9480 } 9481 /** 9482 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection} 9483 * 9484 * <pre> 9485 ** 9486 * This section records information about under-construction files for 9487 * reconstructing the lease map. 9488 * NAME: FILES_UNDERCONSTRUCTION 9489 * </pre> 9490 */ 9491 public static final class Builder extends 9492 com.google.protobuf.GeneratedMessage.Builder<Builder> 9493 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSectionOrBuilder { 9494 public static final com.google.protobuf.Descriptors.Descriptor 9495 getDescriptor() { 9496 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 9497 } 9498 9499 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9500 internalGetFieldAccessorTable() { 9501 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable 9502 .ensureFieldAccessorsInitialized( 9503 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.Builder.class); 9504 } 9505 9506 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.newBuilder() 9507 private Builder() { 9508 maybeForceBuilderInitialization(); 9509 } 9510 9511 private Builder( 9512 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9513 super(parent); 9514 maybeForceBuilderInitialization(); 9515 } 9516 private void maybeForceBuilderInitialization() { 9517 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 9518 } 9519 } 9520 private static Builder create() { 9521 return new Builder(); 9522 } 9523 9524 public Builder clear() { 9525 super.clear(); 9526 return this; 9527 } 9528 9529 public Builder clone() { 9530 return create().mergeFrom(buildPartial()); 9531 } 9532 9533 public com.google.protobuf.Descriptors.Descriptor 9534 getDescriptorForType() { 9535 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 9536 } 9537 9538 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection getDefaultInstanceForType() { 9539 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.getDefaultInstance(); 9540 } 9541 9542 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection build() { 9543 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection result = buildPartial(); 9544 if (!result.isInitialized()) { 9545 throw newUninitializedMessageException(result); 9546 } 9547 return result; 9548 } 9549 9550 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection buildPartial() { 9551 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection(this); 9552 onBuilt(); 9553 return result; 9554 } 9555 9556 public Builder mergeFrom(com.google.protobuf.Message other) { 9557 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection) { 9558 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection)other); 9559 } else { 9560 super.mergeFrom(other); 9561 return this; 9562 } 9563 } 9564 9565 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection other) { 9566 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.getDefaultInstance()) return this; 9567 this.mergeUnknownFields(other.getUnknownFields()); 9568 return this; 9569 } 9570 9571 public final boolean isInitialized() { 9572 return true; 9573 } 9574 9575 public Builder mergeFrom( 9576 com.google.protobuf.CodedInputStream input, 9577 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9578 throws java.io.IOException { 9579 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parsedMessage = null; 9580 try { 9581 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 9582 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9583 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection) e.getUnfinishedMessage(); 9584 throw e; 9585 } finally { 9586 if (parsedMessage != null) { 9587 mergeFrom(parsedMessage); 9588 } 9589 } 9590 return this; 9591 } 9592 9593 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection) 9594 } 9595 9596 static { 9597 defaultInstance = new FilesUnderConstructionSection(true); 9598 defaultInstance.initFields(); 9599 } 9600 9601 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection) 9602 } 9603 9604 public interface INodeDirectorySectionOrBuilder 9605 extends com.google.protobuf.MessageOrBuilder { 9606 } 9607 /** 9608 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection} 9609 * 9610 * <pre> 9611 ** 9612 * This section records the children of each directories 9613 * NAME: INODE_DIR 9614 * </pre> 9615 */ 9616 public static final class INodeDirectorySection extends 9617 com.google.protobuf.GeneratedMessage 9618 implements INodeDirectorySectionOrBuilder { 9619 // Use INodeDirectorySection.newBuilder() to construct. 9620 private INodeDirectorySection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 9621 super(builder); 9622 this.unknownFields = builder.getUnknownFields(); 9623 } 9624 private INodeDirectorySection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 9625 9626 private static final INodeDirectorySection defaultInstance; 9627 public static INodeDirectorySection getDefaultInstance() { 9628 return defaultInstance; 9629 } 9630 9631 public INodeDirectorySection getDefaultInstanceForType() { 9632 return defaultInstance; 9633 } 9634 9635 private final com.google.protobuf.UnknownFieldSet unknownFields; 9636 @java.lang.Override 9637 public final com.google.protobuf.UnknownFieldSet 9638 getUnknownFields() { 9639 return this.unknownFields; 9640 } 9641 private INodeDirectorySection( 9642 com.google.protobuf.CodedInputStream input, 9643 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9644 throws com.google.protobuf.InvalidProtocolBufferException { 9645 initFields(); 9646 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 9647 com.google.protobuf.UnknownFieldSet.newBuilder(); 9648 try { 9649 boolean done = false; 9650 while (!done) { 9651 int tag = input.readTag(); 9652 switch (tag) { 9653 case 0: 9654 done = true; 9655 break; 9656 default: { 9657 if (!parseUnknownField(input, unknownFields, 9658 extensionRegistry, tag)) { 9659 done = true; 9660 } 9661 break; 9662 } 9663 } 9664 } 9665 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9666 throw e.setUnfinishedMessage(this); 9667 } catch (java.io.IOException e) { 9668 throw new com.google.protobuf.InvalidProtocolBufferException( 9669 e.getMessage()).setUnfinishedMessage(this); 9670 } finally { 9671 this.unknownFields = unknownFields.build(); 9672 makeExtensionsImmutable(); 9673 } 9674 } 9675 public static final com.google.protobuf.Descriptors.Descriptor 9676 getDescriptor() { 9677 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 9678 } 9679 9680 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9681 internalGetFieldAccessorTable() { 9682 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable 9683 .ensureFieldAccessorsInitialized( 9684 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.Builder.class); 9685 } 9686 9687 public static com.google.protobuf.Parser<INodeDirectorySection> PARSER = 9688 new com.google.protobuf.AbstractParser<INodeDirectorySection>() { 9689 public INodeDirectorySection parsePartialFrom( 9690 com.google.protobuf.CodedInputStream input, 9691 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9692 throws com.google.protobuf.InvalidProtocolBufferException { 9693 return new INodeDirectorySection(input, extensionRegistry); 9694 } 9695 }; 9696 9697 @java.lang.Override 9698 public com.google.protobuf.Parser<INodeDirectorySection> getParserForType() { 9699 return PARSER; 9700 } 9701 9702 public interface DirEntryOrBuilder 9703 extends com.google.protobuf.MessageOrBuilder { 9704 9705 // optional uint64 parent = 1; 9706 /** 9707 * <code>optional uint64 parent = 1;</code> 9708 */ 9709 boolean hasParent(); 9710 /** 9711 * <code>optional uint64 parent = 1;</code> 9712 */ 9713 long getParent(); 9714 9715 // repeated uint64 children = 2 [packed = true]; 9716 /** 9717 * <code>repeated uint64 children = 2 [packed = true];</code> 9718 * 9719 * <pre> 9720 * children that are not reference nodes 9721 * </pre> 9722 */ 9723 java.util.List<java.lang.Long> getChildrenList(); 9724 /** 9725 * <code>repeated uint64 children = 2 [packed = true];</code> 9726 * 9727 * <pre> 9728 * children that are not reference nodes 9729 * </pre> 9730 */ 9731 int getChildrenCount(); 9732 /** 9733 * <code>repeated uint64 children = 2 [packed = true];</code> 9734 * 9735 * <pre> 9736 * children that are not reference nodes 9737 * </pre> 9738 */ 9739 long getChildren(int index); 9740 9741 // repeated uint32 refChildren = 3 [packed = true]; 9742 /** 9743 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9744 * 9745 * <pre> 9746 * children that are reference nodes, each element is a reference node id 9747 * </pre> 9748 */ 9749 java.util.List<java.lang.Integer> getRefChildrenList(); 9750 /** 9751 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9752 * 9753 * <pre> 9754 * children that are reference nodes, each element is a reference node id 9755 * </pre> 9756 */ 9757 int getRefChildrenCount(); 9758 /** 9759 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9760 * 9761 * <pre> 9762 * children that are reference nodes, each element is a reference node id 9763 * </pre> 9764 */ 9765 int getRefChildren(int index); 9766 } 9767 /** 9768 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry} 9769 * 9770 * <pre> 9771 ** 9772 * A single DirEntry needs to fit in the default PB max message size of 9773 * 64MB. Please be careful when adding more fields to a DirEntry! 9774 * </pre> 9775 */ 9776 public static final class DirEntry extends 9777 com.google.protobuf.GeneratedMessage 9778 implements DirEntryOrBuilder { 9779 // Use DirEntry.newBuilder() to construct. 9780 private DirEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 9781 super(builder); 9782 this.unknownFields = builder.getUnknownFields(); 9783 } 9784 private DirEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 9785 9786 private static final DirEntry defaultInstance; 9787 public static DirEntry getDefaultInstance() { 9788 return defaultInstance; 9789 } 9790 9791 public DirEntry getDefaultInstanceForType() { 9792 return defaultInstance; 9793 } 9794 9795 private final com.google.protobuf.UnknownFieldSet unknownFields; 9796 @java.lang.Override 9797 public final com.google.protobuf.UnknownFieldSet 9798 getUnknownFields() { 9799 return this.unknownFields; 9800 } 9801 private DirEntry( 9802 com.google.protobuf.CodedInputStream input, 9803 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9804 throws com.google.protobuf.InvalidProtocolBufferException { 9805 initFields(); 9806 int mutable_bitField0_ = 0; 9807 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 9808 com.google.protobuf.UnknownFieldSet.newBuilder(); 9809 try { 9810 boolean done = false; 9811 while (!done) { 9812 int tag = input.readTag(); 9813 switch (tag) { 9814 case 0: 9815 done = true; 9816 break; 9817 default: { 9818 if (!parseUnknownField(input, unknownFields, 9819 extensionRegistry, tag)) { 9820 done = true; 9821 } 9822 break; 9823 } 9824 case 8: { 9825 bitField0_ |= 0x00000001; 9826 parent_ = input.readUInt64(); 9827 break; 9828 } 9829 case 16: { 9830 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 9831 children_ = new java.util.ArrayList<java.lang.Long>(); 9832 mutable_bitField0_ |= 0x00000002; 9833 } 9834 children_.add(input.readUInt64()); 9835 break; 9836 } 9837 case 18: { 9838 int length = input.readRawVarint32(); 9839 int limit = input.pushLimit(length); 9840 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) { 9841 children_ = new java.util.ArrayList<java.lang.Long>(); 9842 mutable_bitField0_ |= 0x00000002; 9843 } 9844 while (input.getBytesUntilLimit() > 0) { 9845 children_.add(input.readUInt64()); 9846 } 9847 input.popLimit(limit); 9848 break; 9849 } 9850 case 24: { 9851 if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { 9852 refChildren_ = new java.util.ArrayList<java.lang.Integer>(); 9853 mutable_bitField0_ |= 0x00000004; 9854 } 9855 refChildren_.add(input.readUInt32()); 9856 break; 9857 } 9858 case 26: { 9859 int length = input.readRawVarint32(); 9860 int limit = input.pushLimit(length); 9861 if (!((mutable_bitField0_ & 0x00000004) == 0x00000004) && input.getBytesUntilLimit() > 0) { 9862 refChildren_ = new java.util.ArrayList<java.lang.Integer>(); 9863 mutable_bitField0_ |= 0x00000004; 9864 } 9865 while (input.getBytesUntilLimit() > 0) { 9866 refChildren_.add(input.readUInt32()); 9867 } 9868 input.popLimit(limit); 9869 break; 9870 } 9871 } 9872 } 9873 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9874 throw e.setUnfinishedMessage(this); 9875 } catch (java.io.IOException e) { 9876 throw new com.google.protobuf.InvalidProtocolBufferException( 9877 e.getMessage()).setUnfinishedMessage(this); 9878 } finally { 9879 if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 9880 children_ = java.util.Collections.unmodifiableList(children_); 9881 } 9882 if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { 9883 refChildren_ = java.util.Collections.unmodifiableList(refChildren_); 9884 } 9885 this.unknownFields = unknownFields.build(); 9886 makeExtensionsImmutable(); 9887 } 9888 } 9889 public static final com.google.protobuf.Descriptors.Descriptor 9890 getDescriptor() { 9891 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 9892 } 9893 9894 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9895 internalGetFieldAccessorTable() { 9896 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable 9897 .ensureFieldAccessorsInitialized( 9898 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.Builder.class); 9899 } 9900 9901 public static com.google.protobuf.Parser<DirEntry> PARSER = 9902 new com.google.protobuf.AbstractParser<DirEntry>() { 9903 public DirEntry parsePartialFrom( 9904 com.google.protobuf.CodedInputStream input, 9905 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9906 throws com.google.protobuf.InvalidProtocolBufferException { 9907 return new DirEntry(input, extensionRegistry); 9908 } 9909 }; 9910 9911 @java.lang.Override 9912 public com.google.protobuf.Parser<DirEntry> getParserForType() { 9913 return PARSER; 9914 } 9915 9916 private int bitField0_; 9917 // optional uint64 parent = 1; 9918 public static final int PARENT_FIELD_NUMBER = 1; 9919 private long parent_; 9920 /** 9921 * <code>optional uint64 parent = 1;</code> 9922 */ 9923 public boolean hasParent() { 9924 return ((bitField0_ & 0x00000001) == 0x00000001); 9925 } 9926 /** 9927 * <code>optional uint64 parent = 1;</code> 9928 */ 9929 public long getParent() { 9930 return parent_; 9931 } 9932 9933 // repeated uint64 children = 2 [packed = true]; 9934 public static final int CHILDREN_FIELD_NUMBER = 2; 9935 private java.util.List<java.lang.Long> children_; 9936 /** 9937 * <code>repeated uint64 children = 2 [packed = true];</code> 9938 * 9939 * <pre> 9940 * children that are not reference nodes 9941 * </pre> 9942 */ 9943 public java.util.List<java.lang.Long> 9944 getChildrenList() { 9945 return children_; 9946 } 9947 /** 9948 * <code>repeated uint64 children = 2 [packed = true];</code> 9949 * 9950 * <pre> 9951 * children that are not reference nodes 9952 * </pre> 9953 */ 9954 public int getChildrenCount() { 9955 return children_.size(); 9956 } 9957 /** 9958 * <code>repeated uint64 children = 2 [packed = true];</code> 9959 * 9960 * <pre> 9961 * children that are not reference nodes 9962 * </pre> 9963 */ 9964 public long getChildren(int index) { 9965 return children_.get(index); 9966 } 9967 private int childrenMemoizedSerializedSize = -1; 9968 9969 // repeated uint32 refChildren = 3 [packed = true]; 9970 public static final int REFCHILDREN_FIELD_NUMBER = 3; 9971 private java.util.List<java.lang.Integer> refChildren_; 9972 /** 9973 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9974 * 9975 * <pre> 9976 * children that are reference nodes, each element is a reference node id 9977 * </pre> 9978 */ 9979 public java.util.List<java.lang.Integer> 9980 getRefChildrenList() { 9981 return refChildren_; 9982 } 9983 /** 9984 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9985 * 9986 * <pre> 9987 * children that are reference nodes, each element is a reference node id 9988 * </pre> 9989 */ 9990 public int getRefChildrenCount() { 9991 return refChildren_.size(); 9992 } 9993 /** 9994 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 9995 * 9996 * <pre> 9997 * children that are reference nodes, each element is a reference node id 9998 * </pre> 9999 */ 10000 public int getRefChildren(int index) { 10001 return refChildren_.get(index); 10002 } 10003 private int refChildrenMemoizedSerializedSize = -1; 10004 10005 private void initFields() { 10006 parent_ = 0L; 10007 children_ = java.util.Collections.emptyList(); 10008 refChildren_ = java.util.Collections.emptyList(); 10009 } 10010 private byte memoizedIsInitialized = -1; 10011 public final boolean isInitialized() { 10012 byte isInitialized = memoizedIsInitialized; 10013 if (isInitialized != -1) return isInitialized == 1; 10014 10015 memoizedIsInitialized = 1; 10016 return true; 10017 } 10018 10019 public void writeTo(com.google.protobuf.CodedOutputStream output) 10020 throws java.io.IOException { 10021 getSerializedSize(); 10022 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10023 output.writeUInt64(1, parent_); 10024 } 10025 if (getChildrenList().size() > 0) { 10026 output.writeRawVarint32(18); 10027 output.writeRawVarint32(childrenMemoizedSerializedSize); 10028 } 10029 for (int i = 0; i < children_.size(); i++) { 10030 output.writeUInt64NoTag(children_.get(i)); 10031 } 10032 if (getRefChildrenList().size() > 0) { 10033 output.writeRawVarint32(26); 10034 output.writeRawVarint32(refChildrenMemoizedSerializedSize); 10035 } 10036 for (int i = 0; i < refChildren_.size(); i++) { 10037 output.writeUInt32NoTag(refChildren_.get(i)); 10038 } 10039 getUnknownFields().writeTo(output); 10040 } 10041 10042 private int memoizedSerializedSize = -1; 10043 public int getSerializedSize() { 10044 int size = memoizedSerializedSize; 10045 if (size != -1) return size; 10046 10047 size = 0; 10048 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10049 size += com.google.protobuf.CodedOutputStream 10050 .computeUInt64Size(1, parent_); 10051 } 10052 { 10053 int dataSize = 0; 10054 for (int i = 0; i < children_.size(); i++) { 10055 dataSize += com.google.protobuf.CodedOutputStream 10056 .computeUInt64SizeNoTag(children_.get(i)); 10057 } 10058 size += dataSize; 10059 if (!getChildrenList().isEmpty()) { 10060 size += 1; 10061 size += com.google.protobuf.CodedOutputStream 10062 .computeInt32SizeNoTag(dataSize); 10063 } 10064 childrenMemoizedSerializedSize = dataSize; 10065 } 10066 { 10067 int dataSize = 0; 10068 for (int i = 0; i < refChildren_.size(); i++) { 10069 dataSize += com.google.protobuf.CodedOutputStream 10070 .computeUInt32SizeNoTag(refChildren_.get(i)); 10071 } 10072 size += dataSize; 10073 if (!getRefChildrenList().isEmpty()) { 10074 size += 1; 10075 size += com.google.protobuf.CodedOutputStream 10076 .computeInt32SizeNoTag(dataSize); 10077 } 10078 refChildrenMemoizedSerializedSize = dataSize; 10079 } 10080 size += getUnknownFields().getSerializedSize(); 10081 memoizedSerializedSize = size; 10082 return size; 10083 } 10084 10085 private static final long serialVersionUID = 0L; 10086 @java.lang.Override 10087 protected java.lang.Object writeReplace() 10088 throws java.io.ObjectStreamException { 10089 return super.writeReplace(); 10090 } 10091 10092 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10093 com.google.protobuf.ByteString data) 10094 throws com.google.protobuf.InvalidProtocolBufferException { 10095 return PARSER.parseFrom(data); 10096 } 10097 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10098 com.google.protobuf.ByteString data, 10099 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10100 throws com.google.protobuf.InvalidProtocolBufferException { 10101 return PARSER.parseFrom(data, extensionRegistry); 10102 } 10103 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom(byte[] data) 10104 throws com.google.protobuf.InvalidProtocolBufferException { 10105 return PARSER.parseFrom(data); 10106 } 10107 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10108 byte[] data, 10109 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10110 throws com.google.protobuf.InvalidProtocolBufferException { 10111 return PARSER.parseFrom(data, extensionRegistry); 10112 } 10113 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom(java.io.InputStream input) 10114 throws java.io.IOException { 10115 return PARSER.parseFrom(input); 10116 } 10117 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10118 java.io.InputStream input, 10119 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10120 throws java.io.IOException { 10121 return PARSER.parseFrom(input, extensionRegistry); 10122 } 10123 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseDelimitedFrom(java.io.InputStream input) 10124 throws java.io.IOException { 10125 return PARSER.parseDelimitedFrom(input); 10126 } 10127 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseDelimitedFrom( 10128 java.io.InputStream input, 10129 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10130 throws java.io.IOException { 10131 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10132 } 10133 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10134 com.google.protobuf.CodedInputStream input) 10135 throws java.io.IOException { 10136 return PARSER.parseFrom(input); 10137 } 10138 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 10139 com.google.protobuf.CodedInputStream input, 10140 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10141 throws java.io.IOException { 10142 return PARSER.parseFrom(input, extensionRegistry); 10143 } 10144 10145 public static Builder newBuilder() { return Builder.create(); } 10146 public Builder newBuilderForType() { return newBuilder(); } 10147 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry prototype) { 10148 return newBuilder().mergeFrom(prototype); 10149 } 10150 public Builder toBuilder() { return newBuilder(this); } 10151 10152 @java.lang.Override 10153 protected Builder newBuilderForType( 10154 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10155 Builder builder = new Builder(parent); 10156 return builder; 10157 } 10158 /** 10159 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry} 10160 * 10161 * <pre> 10162 ** 10163 * A single DirEntry needs to fit in the default PB max message size of 10164 * 64MB. Please be careful when adding more fields to a DirEntry! 10165 * </pre> 10166 */ 10167 public static final class Builder extends 10168 com.google.protobuf.GeneratedMessage.Builder<Builder> 10169 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntryOrBuilder { 10170 public static final com.google.protobuf.Descriptors.Descriptor 10171 getDescriptor() { 10172 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 10173 } 10174 10175 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10176 internalGetFieldAccessorTable() { 10177 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable 10178 .ensureFieldAccessorsInitialized( 10179 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.Builder.class); 10180 } 10181 10182 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.newBuilder() 10183 private Builder() { 10184 maybeForceBuilderInitialization(); 10185 } 10186 10187 private Builder( 10188 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10189 super(parent); 10190 maybeForceBuilderInitialization(); 10191 } 10192 private void maybeForceBuilderInitialization() { 10193 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10194 } 10195 } 10196 private static Builder create() { 10197 return new Builder(); 10198 } 10199 10200 public Builder clear() { 10201 super.clear(); 10202 parent_ = 0L; 10203 bitField0_ = (bitField0_ & ~0x00000001); 10204 children_ = java.util.Collections.emptyList(); 10205 bitField0_ = (bitField0_ & ~0x00000002); 10206 refChildren_ = java.util.Collections.emptyList(); 10207 bitField0_ = (bitField0_ & ~0x00000004); 10208 return this; 10209 } 10210 10211 public Builder clone() { 10212 return create().mergeFrom(buildPartial()); 10213 } 10214 10215 public com.google.protobuf.Descriptors.Descriptor 10216 getDescriptorForType() { 10217 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 10218 } 10219 10220 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry getDefaultInstanceForType() { 10221 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.getDefaultInstance(); 10222 } 10223 10224 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry build() { 10225 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry result = buildPartial(); 10226 if (!result.isInitialized()) { 10227 throw newUninitializedMessageException(result); 10228 } 10229 return result; 10230 } 10231 10232 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry buildPartial() { 10233 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry(this); 10234 int from_bitField0_ = bitField0_; 10235 int to_bitField0_ = 0; 10236 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 10237 to_bitField0_ |= 0x00000001; 10238 } 10239 result.parent_ = parent_; 10240 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10241 children_ = java.util.Collections.unmodifiableList(children_); 10242 bitField0_ = (bitField0_ & ~0x00000002); 10243 } 10244 result.children_ = children_; 10245 if (((bitField0_ & 0x00000004) == 0x00000004)) { 10246 refChildren_ = java.util.Collections.unmodifiableList(refChildren_); 10247 bitField0_ = (bitField0_ & ~0x00000004); 10248 } 10249 result.refChildren_ = refChildren_; 10250 result.bitField0_ = to_bitField0_; 10251 onBuilt(); 10252 return result; 10253 } 10254 10255 public Builder mergeFrom(com.google.protobuf.Message other) { 10256 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry) { 10257 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry)other); 10258 } else { 10259 super.mergeFrom(other); 10260 return this; 10261 } 10262 } 10263 10264 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry other) { 10265 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.getDefaultInstance()) return this; 10266 if (other.hasParent()) { 10267 setParent(other.getParent()); 10268 } 10269 if (!other.children_.isEmpty()) { 10270 if (children_.isEmpty()) { 10271 children_ = other.children_; 10272 bitField0_ = (bitField0_ & ~0x00000002); 10273 } else { 10274 ensureChildrenIsMutable(); 10275 children_.addAll(other.children_); 10276 } 10277 onChanged(); 10278 } 10279 if (!other.refChildren_.isEmpty()) { 10280 if (refChildren_.isEmpty()) { 10281 refChildren_ = other.refChildren_; 10282 bitField0_ = (bitField0_ & ~0x00000004); 10283 } else { 10284 ensureRefChildrenIsMutable(); 10285 refChildren_.addAll(other.refChildren_); 10286 } 10287 onChanged(); 10288 } 10289 this.mergeUnknownFields(other.getUnknownFields()); 10290 return this; 10291 } 10292 10293 public final boolean isInitialized() { 10294 return true; 10295 } 10296 10297 public Builder mergeFrom( 10298 com.google.protobuf.CodedInputStream input, 10299 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10300 throws java.io.IOException { 10301 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parsedMessage = null; 10302 try { 10303 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 10304 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10305 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry) e.getUnfinishedMessage(); 10306 throw e; 10307 } finally { 10308 if (parsedMessage != null) { 10309 mergeFrom(parsedMessage); 10310 } 10311 } 10312 return this; 10313 } 10314 private int bitField0_; 10315 10316 // optional uint64 parent = 1; 10317 private long parent_ ; 10318 /** 10319 * <code>optional uint64 parent = 1;</code> 10320 */ 10321 public boolean hasParent() { 10322 return ((bitField0_ & 0x00000001) == 0x00000001); 10323 } 10324 /** 10325 * <code>optional uint64 parent = 1;</code> 10326 */ 10327 public long getParent() { 10328 return parent_; 10329 } 10330 /** 10331 * <code>optional uint64 parent = 1;</code> 10332 */ 10333 public Builder setParent(long value) { 10334 bitField0_ |= 0x00000001; 10335 parent_ = value; 10336 onChanged(); 10337 return this; 10338 } 10339 /** 10340 * <code>optional uint64 parent = 1;</code> 10341 */ 10342 public Builder clearParent() { 10343 bitField0_ = (bitField0_ & ~0x00000001); 10344 parent_ = 0L; 10345 onChanged(); 10346 return this; 10347 } 10348 10349 // repeated uint64 children = 2 [packed = true]; 10350 private java.util.List<java.lang.Long> children_ = java.util.Collections.emptyList(); 10351 private void ensureChildrenIsMutable() { 10352 if (!((bitField0_ & 0x00000002) == 0x00000002)) { 10353 children_ = new java.util.ArrayList<java.lang.Long>(children_); 10354 bitField0_ |= 0x00000002; 10355 } 10356 } 10357 /** 10358 * <code>repeated uint64 children = 2 [packed = true];</code> 10359 * 10360 * <pre> 10361 * children that are not reference nodes 10362 * </pre> 10363 */ 10364 public java.util.List<java.lang.Long> 10365 getChildrenList() { 10366 return java.util.Collections.unmodifiableList(children_); 10367 } 10368 /** 10369 * <code>repeated uint64 children = 2 [packed = true];</code> 10370 * 10371 * <pre> 10372 * children that are not reference nodes 10373 * </pre> 10374 */ 10375 public int getChildrenCount() { 10376 return children_.size(); 10377 } 10378 /** 10379 * <code>repeated uint64 children = 2 [packed = true];</code> 10380 * 10381 * <pre> 10382 * children that are not reference nodes 10383 * </pre> 10384 */ 10385 public long getChildren(int index) { 10386 return children_.get(index); 10387 } 10388 /** 10389 * <code>repeated uint64 children = 2 [packed = true];</code> 10390 * 10391 * <pre> 10392 * children that are not reference nodes 10393 * </pre> 10394 */ 10395 public Builder setChildren( 10396 int index, long value) { 10397 ensureChildrenIsMutable(); 10398 children_.set(index, value); 10399 onChanged(); 10400 return this; 10401 } 10402 /** 10403 * <code>repeated uint64 children = 2 [packed = true];</code> 10404 * 10405 * <pre> 10406 * children that are not reference nodes 10407 * </pre> 10408 */ 10409 public Builder addChildren(long value) { 10410 ensureChildrenIsMutable(); 10411 children_.add(value); 10412 onChanged(); 10413 return this; 10414 } 10415 /** 10416 * <code>repeated uint64 children = 2 [packed = true];</code> 10417 * 10418 * <pre> 10419 * children that are not reference nodes 10420 * </pre> 10421 */ 10422 public Builder addAllChildren( 10423 java.lang.Iterable<? extends java.lang.Long> values) { 10424 ensureChildrenIsMutable(); 10425 super.addAll(values, children_); 10426 onChanged(); 10427 return this; 10428 } 10429 /** 10430 * <code>repeated uint64 children = 2 [packed = true];</code> 10431 * 10432 * <pre> 10433 * children that are not reference nodes 10434 * </pre> 10435 */ 10436 public Builder clearChildren() { 10437 children_ = java.util.Collections.emptyList(); 10438 bitField0_ = (bitField0_ & ~0x00000002); 10439 onChanged(); 10440 return this; 10441 } 10442 10443 // repeated uint32 refChildren = 3 [packed = true]; 10444 private java.util.List<java.lang.Integer> refChildren_ = java.util.Collections.emptyList(); 10445 private void ensureRefChildrenIsMutable() { 10446 if (!((bitField0_ & 0x00000004) == 0x00000004)) { 10447 refChildren_ = new java.util.ArrayList<java.lang.Integer>(refChildren_); 10448 bitField0_ |= 0x00000004; 10449 } 10450 } 10451 /** 10452 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10453 * 10454 * <pre> 10455 * children that are reference nodes, each element is a reference node id 10456 * </pre> 10457 */ 10458 public java.util.List<java.lang.Integer> 10459 getRefChildrenList() { 10460 return java.util.Collections.unmodifiableList(refChildren_); 10461 } 10462 /** 10463 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10464 * 10465 * <pre> 10466 * children that are reference nodes, each element is a reference node id 10467 * </pre> 10468 */ 10469 public int getRefChildrenCount() { 10470 return refChildren_.size(); 10471 } 10472 /** 10473 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10474 * 10475 * <pre> 10476 * children that are reference nodes, each element is a reference node id 10477 * </pre> 10478 */ 10479 public int getRefChildren(int index) { 10480 return refChildren_.get(index); 10481 } 10482 /** 10483 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10484 * 10485 * <pre> 10486 * children that are reference nodes, each element is a reference node id 10487 * </pre> 10488 */ 10489 public Builder setRefChildren( 10490 int index, int value) { 10491 ensureRefChildrenIsMutable(); 10492 refChildren_.set(index, value); 10493 onChanged(); 10494 return this; 10495 } 10496 /** 10497 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10498 * 10499 * <pre> 10500 * children that are reference nodes, each element is a reference node id 10501 * </pre> 10502 */ 10503 public Builder addRefChildren(int value) { 10504 ensureRefChildrenIsMutable(); 10505 refChildren_.add(value); 10506 onChanged(); 10507 return this; 10508 } 10509 /** 10510 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10511 * 10512 * <pre> 10513 * children that are reference nodes, each element is a reference node id 10514 * </pre> 10515 */ 10516 public Builder addAllRefChildren( 10517 java.lang.Iterable<? extends java.lang.Integer> values) { 10518 ensureRefChildrenIsMutable(); 10519 super.addAll(values, refChildren_); 10520 onChanged(); 10521 return this; 10522 } 10523 /** 10524 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 10525 * 10526 * <pre> 10527 * children that are reference nodes, each element is a reference node id 10528 * </pre> 10529 */ 10530 public Builder clearRefChildren() { 10531 refChildren_ = java.util.Collections.emptyList(); 10532 bitField0_ = (bitField0_ & ~0x00000004); 10533 onChanged(); 10534 return this; 10535 } 10536 10537 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry) 10538 } 10539 10540 static { 10541 defaultInstance = new DirEntry(true); 10542 defaultInstance.initFields(); 10543 } 10544 10545 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry) 10546 } 10547 10548 private void initFields() { 10549 } 10550 private byte memoizedIsInitialized = -1; 10551 public final boolean isInitialized() { 10552 byte isInitialized = memoizedIsInitialized; 10553 if (isInitialized != -1) return isInitialized == 1; 10554 10555 memoizedIsInitialized = 1; 10556 return true; 10557 } 10558 10559 public void writeTo(com.google.protobuf.CodedOutputStream output) 10560 throws java.io.IOException { 10561 getSerializedSize(); 10562 getUnknownFields().writeTo(output); 10563 } 10564 10565 private int memoizedSerializedSize = -1; 10566 public int getSerializedSize() { 10567 int size = memoizedSerializedSize; 10568 if (size != -1) return size; 10569 10570 size = 0; 10571 size += getUnknownFields().getSerializedSize(); 10572 memoizedSerializedSize = size; 10573 return size; 10574 } 10575 10576 private static final long serialVersionUID = 0L; 10577 @java.lang.Override 10578 protected java.lang.Object writeReplace() 10579 throws java.io.ObjectStreamException { 10580 return super.writeReplace(); 10581 } 10582 10583 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10584 com.google.protobuf.ByteString data) 10585 throws com.google.protobuf.InvalidProtocolBufferException { 10586 return PARSER.parseFrom(data); 10587 } 10588 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10589 com.google.protobuf.ByteString data, 10590 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10591 throws com.google.protobuf.InvalidProtocolBufferException { 10592 return PARSER.parseFrom(data, extensionRegistry); 10593 } 10594 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom(byte[] data) 10595 throws com.google.protobuf.InvalidProtocolBufferException { 10596 return PARSER.parseFrom(data); 10597 } 10598 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10599 byte[] data, 10600 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10601 throws com.google.protobuf.InvalidProtocolBufferException { 10602 return PARSER.parseFrom(data, extensionRegistry); 10603 } 10604 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom(java.io.InputStream input) 10605 throws java.io.IOException { 10606 return PARSER.parseFrom(input); 10607 } 10608 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10609 java.io.InputStream input, 10610 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10611 throws java.io.IOException { 10612 return PARSER.parseFrom(input, extensionRegistry); 10613 } 10614 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseDelimitedFrom(java.io.InputStream input) 10615 throws java.io.IOException { 10616 return PARSER.parseDelimitedFrom(input); 10617 } 10618 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseDelimitedFrom( 10619 java.io.InputStream input, 10620 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10621 throws java.io.IOException { 10622 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10623 } 10624 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10625 com.google.protobuf.CodedInputStream input) 10626 throws java.io.IOException { 10627 return PARSER.parseFrom(input); 10628 } 10629 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 10630 com.google.protobuf.CodedInputStream input, 10631 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10632 throws java.io.IOException { 10633 return PARSER.parseFrom(input, extensionRegistry); 10634 } 10635 10636 public static Builder newBuilder() { return Builder.create(); } 10637 public Builder newBuilderForType() { return newBuilder(); } 10638 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection prototype) { 10639 return newBuilder().mergeFrom(prototype); 10640 } 10641 public Builder toBuilder() { return newBuilder(this); } 10642 10643 @java.lang.Override 10644 protected Builder newBuilderForType( 10645 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10646 Builder builder = new Builder(parent); 10647 return builder; 10648 } 10649 /** 10650 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection} 10651 * 10652 * <pre> 10653 ** 10654 * This section records the children of each directories 10655 * NAME: INODE_DIR 10656 * </pre> 10657 */ 10658 public static final class Builder extends 10659 com.google.protobuf.GeneratedMessage.Builder<Builder> 10660 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySectionOrBuilder { 10661 public static final com.google.protobuf.Descriptors.Descriptor 10662 getDescriptor() { 10663 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 10664 } 10665 10666 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10667 internalGetFieldAccessorTable() { 10668 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable 10669 .ensureFieldAccessorsInitialized( 10670 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.Builder.class); 10671 } 10672 10673 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.newBuilder() 10674 private Builder() { 10675 maybeForceBuilderInitialization(); 10676 } 10677 10678 private Builder( 10679 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10680 super(parent); 10681 maybeForceBuilderInitialization(); 10682 } 10683 private void maybeForceBuilderInitialization() { 10684 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10685 } 10686 } 10687 private static Builder create() { 10688 return new Builder(); 10689 } 10690 10691 public Builder clear() { 10692 super.clear(); 10693 return this; 10694 } 10695 10696 public Builder clone() { 10697 return create().mergeFrom(buildPartial()); 10698 } 10699 10700 public com.google.protobuf.Descriptors.Descriptor 10701 getDescriptorForType() { 10702 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 10703 } 10704 10705 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection getDefaultInstanceForType() { 10706 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.getDefaultInstance(); 10707 } 10708 10709 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection build() { 10710 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection result = buildPartial(); 10711 if (!result.isInitialized()) { 10712 throw newUninitializedMessageException(result); 10713 } 10714 return result; 10715 } 10716 10717 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection buildPartial() { 10718 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection(this); 10719 onBuilt(); 10720 return result; 10721 } 10722 10723 public Builder mergeFrom(com.google.protobuf.Message other) { 10724 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection) { 10725 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection)other); 10726 } else { 10727 super.mergeFrom(other); 10728 return this; 10729 } 10730 } 10731 10732 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection other) { 10733 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.getDefaultInstance()) return this; 10734 this.mergeUnknownFields(other.getUnknownFields()); 10735 return this; 10736 } 10737 10738 public final boolean isInitialized() { 10739 return true; 10740 } 10741 10742 public Builder mergeFrom( 10743 com.google.protobuf.CodedInputStream input, 10744 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10745 throws java.io.IOException { 10746 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parsedMessage = null; 10747 try { 10748 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 10749 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10750 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection) e.getUnfinishedMessage(); 10751 throw e; 10752 } finally { 10753 if (parsedMessage != null) { 10754 mergeFrom(parsedMessage); 10755 } 10756 } 10757 return this; 10758 } 10759 10760 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeDirectorySection) 10761 } 10762 10763 static { 10764 defaultInstance = new INodeDirectorySection(true); 10765 defaultInstance.initFields(); 10766 } 10767 10768 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeDirectorySection) 10769 } 10770 10771 public interface INodeReferenceSectionOrBuilder 10772 extends com.google.protobuf.MessageOrBuilder { 10773 } 10774 /** 10775 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection} 10776 */ 10777 public static final class INodeReferenceSection extends 10778 com.google.protobuf.GeneratedMessage 10779 implements INodeReferenceSectionOrBuilder { 10780 // Use INodeReferenceSection.newBuilder() to construct. 10781 private INodeReferenceSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 10782 super(builder); 10783 this.unknownFields = builder.getUnknownFields(); 10784 } 10785 private INodeReferenceSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 10786 10787 private static final INodeReferenceSection defaultInstance; 10788 public static INodeReferenceSection getDefaultInstance() { 10789 return defaultInstance; 10790 } 10791 10792 public INodeReferenceSection getDefaultInstanceForType() { 10793 return defaultInstance; 10794 } 10795 10796 private final com.google.protobuf.UnknownFieldSet unknownFields; 10797 @java.lang.Override 10798 public final com.google.protobuf.UnknownFieldSet 10799 getUnknownFields() { 10800 return this.unknownFields; 10801 } 10802 private INodeReferenceSection( 10803 com.google.protobuf.CodedInputStream input, 10804 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10805 throws com.google.protobuf.InvalidProtocolBufferException { 10806 initFields(); 10807 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 10808 com.google.protobuf.UnknownFieldSet.newBuilder(); 10809 try { 10810 boolean done = false; 10811 while (!done) { 10812 int tag = input.readTag(); 10813 switch (tag) { 10814 case 0: 10815 done = true; 10816 break; 10817 default: { 10818 if (!parseUnknownField(input, unknownFields, 10819 extensionRegistry, tag)) { 10820 done = true; 10821 } 10822 break; 10823 } 10824 } 10825 } 10826 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10827 throw e.setUnfinishedMessage(this); 10828 } catch (java.io.IOException e) { 10829 throw new com.google.protobuf.InvalidProtocolBufferException( 10830 e.getMessage()).setUnfinishedMessage(this); 10831 } finally { 10832 this.unknownFields = unknownFields.build(); 10833 makeExtensionsImmutable(); 10834 } 10835 } 10836 public static final com.google.protobuf.Descriptors.Descriptor 10837 getDescriptor() { 10838 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 10839 } 10840 10841 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10842 internalGetFieldAccessorTable() { 10843 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable 10844 .ensureFieldAccessorsInitialized( 10845 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.Builder.class); 10846 } 10847 10848 public static com.google.protobuf.Parser<INodeReferenceSection> PARSER = 10849 new com.google.protobuf.AbstractParser<INodeReferenceSection>() { 10850 public INodeReferenceSection parsePartialFrom( 10851 com.google.protobuf.CodedInputStream input, 10852 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10853 throws com.google.protobuf.InvalidProtocolBufferException { 10854 return new INodeReferenceSection(input, extensionRegistry); 10855 } 10856 }; 10857 10858 @java.lang.Override 10859 public com.google.protobuf.Parser<INodeReferenceSection> getParserForType() { 10860 return PARSER; 10861 } 10862 10863 public interface INodeReferenceOrBuilder 10864 extends com.google.protobuf.MessageOrBuilder { 10865 10866 // optional uint64 referredId = 1; 10867 /** 10868 * <code>optional uint64 referredId = 1;</code> 10869 * 10870 * <pre> 10871 * id of the referred inode 10872 * </pre> 10873 */ 10874 boolean hasReferredId(); 10875 /** 10876 * <code>optional uint64 referredId = 1;</code> 10877 * 10878 * <pre> 10879 * id of the referred inode 10880 * </pre> 10881 */ 10882 long getReferredId(); 10883 10884 // optional bytes name = 2; 10885 /** 10886 * <code>optional bytes name = 2;</code> 10887 * 10888 * <pre> 10889 * local name recorded in WithName 10890 * </pre> 10891 */ 10892 boolean hasName(); 10893 /** 10894 * <code>optional bytes name = 2;</code> 10895 * 10896 * <pre> 10897 * local name recorded in WithName 10898 * </pre> 10899 */ 10900 com.google.protobuf.ByteString getName(); 10901 10902 // optional uint32 dstSnapshotId = 3; 10903 /** 10904 * <code>optional uint32 dstSnapshotId = 3;</code> 10905 * 10906 * <pre> 10907 * recorded in DstReference 10908 * </pre> 10909 */ 10910 boolean hasDstSnapshotId(); 10911 /** 10912 * <code>optional uint32 dstSnapshotId = 3;</code> 10913 * 10914 * <pre> 10915 * recorded in DstReference 10916 * </pre> 10917 */ 10918 int getDstSnapshotId(); 10919 10920 // optional uint32 lastSnapshotId = 4; 10921 /** 10922 * <code>optional uint32 lastSnapshotId = 4;</code> 10923 * 10924 * <pre> 10925 * recorded in WithName 10926 * </pre> 10927 */ 10928 boolean hasLastSnapshotId(); 10929 /** 10930 * <code>optional uint32 lastSnapshotId = 4;</code> 10931 * 10932 * <pre> 10933 * recorded in WithName 10934 * </pre> 10935 */ 10936 int getLastSnapshotId(); 10937 } 10938 /** 10939 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference} 10940 */ 10941 public static final class INodeReference extends 10942 com.google.protobuf.GeneratedMessage 10943 implements INodeReferenceOrBuilder { 10944 // Use INodeReference.newBuilder() to construct. 10945 private INodeReference(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 10946 super(builder); 10947 this.unknownFields = builder.getUnknownFields(); 10948 } 10949 private INodeReference(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 10950 10951 private static final INodeReference defaultInstance; 10952 public static INodeReference getDefaultInstance() { 10953 return defaultInstance; 10954 } 10955 10956 public INodeReference getDefaultInstanceForType() { 10957 return defaultInstance; 10958 } 10959 10960 private final com.google.protobuf.UnknownFieldSet unknownFields; 10961 @java.lang.Override 10962 public final com.google.protobuf.UnknownFieldSet 10963 getUnknownFields() { 10964 return this.unknownFields; 10965 } 10966 private INodeReference( 10967 com.google.protobuf.CodedInputStream input, 10968 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10969 throws com.google.protobuf.InvalidProtocolBufferException { 10970 initFields(); 10971 int mutable_bitField0_ = 0; 10972 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 10973 com.google.protobuf.UnknownFieldSet.newBuilder(); 10974 try { 10975 boolean done = false; 10976 while (!done) { 10977 int tag = input.readTag(); 10978 switch (tag) { 10979 case 0: 10980 done = true; 10981 break; 10982 default: { 10983 if (!parseUnknownField(input, unknownFields, 10984 extensionRegistry, tag)) { 10985 done = true; 10986 } 10987 break; 10988 } 10989 case 8: { 10990 bitField0_ |= 0x00000001; 10991 referredId_ = input.readUInt64(); 10992 break; 10993 } 10994 case 18: { 10995 bitField0_ |= 0x00000002; 10996 name_ = input.readBytes(); 10997 break; 10998 } 10999 case 24: { 11000 bitField0_ |= 0x00000004; 11001 dstSnapshotId_ = input.readUInt32(); 11002 break; 11003 } 11004 case 32: { 11005 bitField0_ |= 0x00000008; 11006 lastSnapshotId_ = input.readUInt32(); 11007 break; 11008 } 11009 } 11010 } 11011 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11012 throw e.setUnfinishedMessage(this); 11013 } catch (java.io.IOException e) { 11014 throw new com.google.protobuf.InvalidProtocolBufferException( 11015 e.getMessage()).setUnfinishedMessage(this); 11016 } finally { 11017 this.unknownFields = unknownFields.build(); 11018 makeExtensionsImmutable(); 11019 } 11020 } 11021 public static final com.google.protobuf.Descriptors.Descriptor 11022 getDescriptor() { 11023 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 11024 } 11025 11026 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11027 internalGetFieldAccessorTable() { 11028 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable 11029 .ensureFieldAccessorsInitialized( 11030 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.Builder.class); 11031 } 11032 11033 public static com.google.protobuf.Parser<INodeReference> PARSER = 11034 new com.google.protobuf.AbstractParser<INodeReference>() { 11035 public INodeReference parsePartialFrom( 11036 com.google.protobuf.CodedInputStream input, 11037 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11038 throws com.google.protobuf.InvalidProtocolBufferException { 11039 return new INodeReference(input, extensionRegistry); 11040 } 11041 }; 11042 11043 @java.lang.Override 11044 public com.google.protobuf.Parser<INodeReference> getParserForType() { 11045 return PARSER; 11046 } 11047 11048 private int bitField0_; 11049 // optional uint64 referredId = 1; 11050 public static final int REFERREDID_FIELD_NUMBER = 1; 11051 private long referredId_; 11052 /** 11053 * <code>optional uint64 referredId = 1;</code> 11054 * 11055 * <pre> 11056 * id of the referred inode 11057 * </pre> 11058 */ 11059 public boolean hasReferredId() { 11060 return ((bitField0_ & 0x00000001) == 0x00000001); 11061 } 11062 /** 11063 * <code>optional uint64 referredId = 1;</code> 11064 * 11065 * <pre> 11066 * id of the referred inode 11067 * </pre> 11068 */ 11069 public long getReferredId() { 11070 return referredId_; 11071 } 11072 11073 // optional bytes name = 2; 11074 public static final int NAME_FIELD_NUMBER = 2; 11075 private com.google.protobuf.ByteString name_; 11076 /** 11077 * <code>optional bytes name = 2;</code> 11078 * 11079 * <pre> 11080 * local name recorded in WithName 11081 * </pre> 11082 */ 11083 public boolean hasName() { 11084 return ((bitField0_ & 0x00000002) == 0x00000002); 11085 } 11086 /** 11087 * <code>optional bytes name = 2;</code> 11088 * 11089 * <pre> 11090 * local name recorded in WithName 11091 * </pre> 11092 */ 11093 public com.google.protobuf.ByteString getName() { 11094 return name_; 11095 } 11096 11097 // optional uint32 dstSnapshotId = 3; 11098 public static final int DSTSNAPSHOTID_FIELD_NUMBER = 3; 11099 private int dstSnapshotId_; 11100 /** 11101 * <code>optional uint32 dstSnapshotId = 3;</code> 11102 * 11103 * <pre> 11104 * recorded in DstReference 11105 * </pre> 11106 */ 11107 public boolean hasDstSnapshotId() { 11108 return ((bitField0_ & 0x00000004) == 0x00000004); 11109 } 11110 /** 11111 * <code>optional uint32 dstSnapshotId = 3;</code> 11112 * 11113 * <pre> 11114 * recorded in DstReference 11115 * </pre> 11116 */ 11117 public int getDstSnapshotId() { 11118 return dstSnapshotId_; 11119 } 11120 11121 // optional uint32 lastSnapshotId = 4; 11122 public static final int LASTSNAPSHOTID_FIELD_NUMBER = 4; 11123 private int lastSnapshotId_; 11124 /** 11125 * <code>optional uint32 lastSnapshotId = 4;</code> 11126 * 11127 * <pre> 11128 * recorded in WithName 11129 * </pre> 11130 */ 11131 public boolean hasLastSnapshotId() { 11132 return ((bitField0_ & 0x00000008) == 0x00000008); 11133 } 11134 /** 11135 * <code>optional uint32 lastSnapshotId = 4;</code> 11136 * 11137 * <pre> 11138 * recorded in WithName 11139 * </pre> 11140 */ 11141 public int getLastSnapshotId() { 11142 return lastSnapshotId_; 11143 } 11144 11145 private void initFields() { 11146 referredId_ = 0L; 11147 name_ = com.google.protobuf.ByteString.EMPTY; 11148 dstSnapshotId_ = 0; 11149 lastSnapshotId_ = 0; 11150 } 11151 private byte memoizedIsInitialized = -1; 11152 public final boolean isInitialized() { 11153 byte isInitialized = memoizedIsInitialized; 11154 if (isInitialized != -1) return isInitialized == 1; 11155 11156 memoizedIsInitialized = 1; 11157 return true; 11158 } 11159 11160 public void writeTo(com.google.protobuf.CodedOutputStream output) 11161 throws java.io.IOException { 11162 getSerializedSize(); 11163 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11164 output.writeUInt64(1, referredId_); 11165 } 11166 if (((bitField0_ & 0x00000002) == 0x00000002)) { 11167 output.writeBytes(2, name_); 11168 } 11169 if (((bitField0_ & 0x00000004) == 0x00000004)) { 11170 output.writeUInt32(3, dstSnapshotId_); 11171 } 11172 if (((bitField0_ & 0x00000008) == 0x00000008)) { 11173 output.writeUInt32(4, lastSnapshotId_); 11174 } 11175 getUnknownFields().writeTo(output); 11176 } 11177 11178 private int memoizedSerializedSize = -1; 11179 public int getSerializedSize() { 11180 int size = memoizedSerializedSize; 11181 if (size != -1) return size; 11182 11183 size = 0; 11184 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11185 size += com.google.protobuf.CodedOutputStream 11186 .computeUInt64Size(1, referredId_); 11187 } 11188 if (((bitField0_ & 0x00000002) == 0x00000002)) { 11189 size += com.google.protobuf.CodedOutputStream 11190 .computeBytesSize(2, name_); 11191 } 11192 if (((bitField0_ & 0x00000004) == 0x00000004)) { 11193 size += com.google.protobuf.CodedOutputStream 11194 .computeUInt32Size(3, dstSnapshotId_); 11195 } 11196 if (((bitField0_ & 0x00000008) == 0x00000008)) { 11197 size += com.google.protobuf.CodedOutputStream 11198 .computeUInt32Size(4, lastSnapshotId_); 11199 } 11200 size += getUnknownFields().getSerializedSize(); 11201 memoizedSerializedSize = size; 11202 return size; 11203 } 11204 11205 private static final long serialVersionUID = 0L; 11206 @java.lang.Override 11207 protected java.lang.Object writeReplace() 11208 throws java.io.ObjectStreamException { 11209 return super.writeReplace(); 11210 } 11211 11212 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11213 com.google.protobuf.ByteString data) 11214 throws com.google.protobuf.InvalidProtocolBufferException { 11215 return PARSER.parseFrom(data); 11216 } 11217 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11218 com.google.protobuf.ByteString data, 11219 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11220 throws com.google.protobuf.InvalidProtocolBufferException { 11221 return PARSER.parseFrom(data, extensionRegistry); 11222 } 11223 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom(byte[] data) 11224 throws com.google.protobuf.InvalidProtocolBufferException { 11225 return PARSER.parseFrom(data); 11226 } 11227 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11228 byte[] data, 11229 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11230 throws com.google.protobuf.InvalidProtocolBufferException { 11231 return PARSER.parseFrom(data, extensionRegistry); 11232 } 11233 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom(java.io.InputStream input) 11234 throws java.io.IOException { 11235 return PARSER.parseFrom(input); 11236 } 11237 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11238 java.io.InputStream input, 11239 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11240 throws java.io.IOException { 11241 return PARSER.parseFrom(input, extensionRegistry); 11242 } 11243 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseDelimitedFrom(java.io.InputStream input) 11244 throws java.io.IOException { 11245 return PARSER.parseDelimitedFrom(input); 11246 } 11247 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseDelimitedFrom( 11248 java.io.InputStream input, 11249 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11250 throws java.io.IOException { 11251 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11252 } 11253 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11254 com.google.protobuf.CodedInputStream input) 11255 throws java.io.IOException { 11256 return PARSER.parseFrom(input); 11257 } 11258 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 11259 com.google.protobuf.CodedInputStream input, 11260 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11261 throws java.io.IOException { 11262 return PARSER.parseFrom(input, extensionRegistry); 11263 } 11264 11265 public static Builder newBuilder() { return Builder.create(); } 11266 public Builder newBuilderForType() { return newBuilder(); } 11267 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference prototype) { 11268 return newBuilder().mergeFrom(prototype); 11269 } 11270 public Builder toBuilder() { return newBuilder(this); } 11271 11272 @java.lang.Override 11273 protected Builder newBuilderForType( 11274 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11275 Builder builder = new Builder(parent); 11276 return builder; 11277 } 11278 /** 11279 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference} 11280 */ 11281 public static final class Builder extends 11282 com.google.protobuf.GeneratedMessage.Builder<Builder> 11283 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReferenceOrBuilder { 11284 public static final com.google.protobuf.Descriptors.Descriptor 11285 getDescriptor() { 11286 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 11287 } 11288 11289 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11290 internalGetFieldAccessorTable() { 11291 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable 11292 .ensureFieldAccessorsInitialized( 11293 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.Builder.class); 11294 } 11295 11296 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.newBuilder() 11297 private Builder() { 11298 maybeForceBuilderInitialization(); 11299 } 11300 11301 private Builder( 11302 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11303 super(parent); 11304 maybeForceBuilderInitialization(); 11305 } 11306 private void maybeForceBuilderInitialization() { 11307 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 11308 } 11309 } 11310 private static Builder create() { 11311 return new Builder(); 11312 } 11313 11314 public Builder clear() { 11315 super.clear(); 11316 referredId_ = 0L; 11317 bitField0_ = (bitField0_ & ~0x00000001); 11318 name_ = com.google.protobuf.ByteString.EMPTY; 11319 bitField0_ = (bitField0_ & ~0x00000002); 11320 dstSnapshotId_ = 0; 11321 bitField0_ = (bitField0_ & ~0x00000004); 11322 lastSnapshotId_ = 0; 11323 bitField0_ = (bitField0_ & ~0x00000008); 11324 return this; 11325 } 11326 11327 public Builder clone() { 11328 return create().mergeFrom(buildPartial()); 11329 } 11330 11331 public com.google.protobuf.Descriptors.Descriptor 11332 getDescriptorForType() { 11333 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 11334 } 11335 11336 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference getDefaultInstanceForType() { 11337 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.getDefaultInstance(); 11338 } 11339 11340 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference build() { 11341 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference result = buildPartial(); 11342 if (!result.isInitialized()) { 11343 throw newUninitializedMessageException(result); 11344 } 11345 return result; 11346 } 11347 11348 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference buildPartial() { 11349 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference(this); 11350 int from_bitField0_ = bitField0_; 11351 int to_bitField0_ = 0; 11352 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 11353 to_bitField0_ |= 0x00000001; 11354 } 11355 result.referredId_ = referredId_; 11356 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 11357 to_bitField0_ |= 0x00000002; 11358 } 11359 result.name_ = name_; 11360 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 11361 to_bitField0_ |= 0x00000004; 11362 } 11363 result.dstSnapshotId_ = dstSnapshotId_; 11364 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 11365 to_bitField0_ |= 0x00000008; 11366 } 11367 result.lastSnapshotId_ = lastSnapshotId_; 11368 result.bitField0_ = to_bitField0_; 11369 onBuilt(); 11370 return result; 11371 } 11372 11373 public Builder mergeFrom(com.google.protobuf.Message other) { 11374 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference) { 11375 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference)other); 11376 } else { 11377 super.mergeFrom(other); 11378 return this; 11379 } 11380 } 11381 11382 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference other) { 11383 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.getDefaultInstance()) return this; 11384 if (other.hasReferredId()) { 11385 setReferredId(other.getReferredId()); 11386 } 11387 if (other.hasName()) { 11388 setName(other.getName()); 11389 } 11390 if (other.hasDstSnapshotId()) { 11391 setDstSnapshotId(other.getDstSnapshotId()); 11392 } 11393 if (other.hasLastSnapshotId()) { 11394 setLastSnapshotId(other.getLastSnapshotId()); 11395 } 11396 this.mergeUnknownFields(other.getUnknownFields()); 11397 return this; 11398 } 11399 11400 public final boolean isInitialized() { 11401 return true; 11402 } 11403 11404 public Builder mergeFrom( 11405 com.google.protobuf.CodedInputStream input, 11406 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11407 throws java.io.IOException { 11408 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parsedMessage = null; 11409 try { 11410 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 11411 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11412 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference) e.getUnfinishedMessage(); 11413 throw e; 11414 } finally { 11415 if (parsedMessage != null) { 11416 mergeFrom(parsedMessage); 11417 } 11418 } 11419 return this; 11420 } 11421 private int bitField0_; 11422 11423 // optional uint64 referredId = 1; 11424 private long referredId_ ; 11425 /** 11426 * <code>optional uint64 referredId = 1;</code> 11427 * 11428 * <pre> 11429 * id of the referred inode 11430 * </pre> 11431 */ 11432 public boolean hasReferredId() { 11433 return ((bitField0_ & 0x00000001) == 0x00000001); 11434 } 11435 /** 11436 * <code>optional uint64 referredId = 1;</code> 11437 * 11438 * <pre> 11439 * id of the referred inode 11440 * </pre> 11441 */ 11442 public long getReferredId() { 11443 return referredId_; 11444 } 11445 /** 11446 * <code>optional uint64 referredId = 1;</code> 11447 * 11448 * <pre> 11449 * id of the referred inode 11450 * </pre> 11451 */ 11452 public Builder setReferredId(long value) { 11453 bitField0_ |= 0x00000001; 11454 referredId_ = value; 11455 onChanged(); 11456 return this; 11457 } 11458 /** 11459 * <code>optional uint64 referredId = 1;</code> 11460 * 11461 * <pre> 11462 * id of the referred inode 11463 * </pre> 11464 */ 11465 public Builder clearReferredId() { 11466 bitField0_ = (bitField0_ & ~0x00000001); 11467 referredId_ = 0L; 11468 onChanged(); 11469 return this; 11470 } 11471 11472 // optional bytes name = 2; 11473 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 11474 /** 11475 * <code>optional bytes name = 2;</code> 11476 * 11477 * <pre> 11478 * local name recorded in WithName 11479 * </pre> 11480 */ 11481 public boolean hasName() { 11482 return ((bitField0_ & 0x00000002) == 0x00000002); 11483 } 11484 /** 11485 * <code>optional bytes name = 2;</code> 11486 * 11487 * <pre> 11488 * local name recorded in WithName 11489 * </pre> 11490 */ 11491 public com.google.protobuf.ByteString getName() { 11492 return name_; 11493 } 11494 /** 11495 * <code>optional bytes name = 2;</code> 11496 * 11497 * <pre> 11498 * local name recorded in WithName 11499 * </pre> 11500 */ 11501 public Builder setName(com.google.protobuf.ByteString value) { 11502 if (value == null) { 11503 throw new NullPointerException(); 11504 } 11505 bitField0_ |= 0x00000002; 11506 name_ = value; 11507 onChanged(); 11508 return this; 11509 } 11510 /** 11511 * <code>optional bytes name = 2;</code> 11512 * 11513 * <pre> 11514 * local name recorded in WithName 11515 * </pre> 11516 */ 11517 public Builder clearName() { 11518 bitField0_ = (bitField0_ & ~0x00000002); 11519 name_ = getDefaultInstance().getName(); 11520 onChanged(); 11521 return this; 11522 } 11523 11524 // optional uint32 dstSnapshotId = 3; 11525 private int dstSnapshotId_ ; 11526 /** 11527 * <code>optional uint32 dstSnapshotId = 3;</code> 11528 * 11529 * <pre> 11530 * recorded in DstReference 11531 * </pre> 11532 */ 11533 public boolean hasDstSnapshotId() { 11534 return ((bitField0_ & 0x00000004) == 0x00000004); 11535 } 11536 /** 11537 * <code>optional uint32 dstSnapshotId = 3;</code> 11538 * 11539 * <pre> 11540 * recorded in DstReference 11541 * </pre> 11542 */ 11543 public int getDstSnapshotId() { 11544 return dstSnapshotId_; 11545 } 11546 /** 11547 * <code>optional uint32 dstSnapshotId = 3;</code> 11548 * 11549 * <pre> 11550 * recorded in DstReference 11551 * </pre> 11552 */ 11553 public Builder setDstSnapshotId(int value) { 11554 bitField0_ |= 0x00000004; 11555 dstSnapshotId_ = value; 11556 onChanged(); 11557 return this; 11558 } 11559 /** 11560 * <code>optional uint32 dstSnapshotId = 3;</code> 11561 * 11562 * <pre> 11563 * recorded in DstReference 11564 * </pre> 11565 */ 11566 public Builder clearDstSnapshotId() { 11567 bitField0_ = (bitField0_ & ~0x00000004); 11568 dstSnapshotId_ = 0; 11569 onChanged(); 11570 return this; 11571 } 11572 11573 // optional uint32 lastSnapshotId = 4; 11574 private int lastSnapshotId_ ; 11575 /** 11576 * <code>optional uint32 lastSnapshotId = 4;</code> 11577 * 11578 * <pre> 11579 * recorded in WithName 11580 * </pre> 11581 */ 11582 public boolean hasLastSnapshotId() { 11583 return ((bitField0_ & 0x00000008) == 0x00000008); 11584 } 11585 /** 11586 * <code>optional uint32 lastSnapshotId = 4;</code> 11587 * 11588 * <pre> 11589 * recorded in WithName 11590 * </pre> 11591 */ 11592 public int getLastSnapshotId() { 11593 return lastSnapshotId_; 11594 } 11595 /** 11596 * <code>optional uint32 lastSnapshotId = 4;</code> 11597 * 11598 * <pre> 11599 * recorded in WithName 11600 * </pre> 11601 */ 11602 public Builder setLastSnapshotId(int value) { 11603 bitField0_ |= 0x00000008; 11604 lastSnapshotId_ = value; 11605 onChanged(); 11606 return this; 11607 } 11608 /** 11609 * <code>optional uint32 lastSnapshotId = 4;</code> 11610 * 11611 * <pre> 11612 * recorded in WithName 11613 * </pre> 11614 */ 11615 public Builder clearLastSnapshotId() { 11616 bitField0_ = (bitField0_ & ~0x00000008); 11617 lastSnapshotId_ = 0; 11618 onChanged(); 11619 return this; 11620 } 11621 11622 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference) 11623 } 11624 11625 static { 11626 defaultInstance = new INodeReference(true); 11627 defaultInstance.initFields(); 11628 } 11629 11630 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference) 11631 } 11632 11633 private void initFields() { 11634 } 11635 private byte memoizedIsInitialized = -1; 11636 public final boolean isInitialized() { 11637 byte isInitialized = memoizedIsInitialized; 11638 if (isInitialized != -1) return isInitialized == 1; 11639 11640 memoizedIsInitialized = 1; 11641 return true; 11642 } 11643 11644 public void writeTo(com.google.protobuf.CodedOutputStream output) 11645 throws java.io.IOException { 11646 getSerializedSize(); 11647 getUnknownFields().writeTo(output); 11648 } 11649 11650 private int memoizedSerializedSize = -1; 11651 public int getSerializedSize() { 11652 int size = memoizedSerializedSize; 11653 if (size != -1) return size; 11654 11655 size = 0; 11656 size += getUnknownFields().getSerializedSize(); 11657 memoizedSerializedSize = size; 11658 return size; 11659 } 11660 11661 private static final long serialVersionUID = 0L; 11662 @java.lang.Override 11663 protected java.lang.Object writeReplace() 11664 throws java.io.ObjectStreamException { 11665 return super.writeReplace(); 11666 } 11667 11668 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11669 com.google.protobuf.ByteString data) 11670 throws com.google.protobuf.InvalidProtocolBufferException { 11671 return PARSER.parseFrom(data); 11672 } 11673 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11674 com.google.protobuf.ByteString data, 11675 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11676 throws com.google.protobuf.InvalidProtocolBufferException { 11677 return PARSER.parseFrom(data, extensionRegistry); 11678 } 11679 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom(byte[] data) 11680 throws com.google.protobuf.InvalidProtocolBufferException { 11681 return PARSER.parseFrom(data); 11682 } 11683 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11684 byte[] data, 11685 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11686 throws com.google.protobuf.InvalidProtocolBufferException { 11687 return PARSER.parseFrom(data, extensionRegistry); 11688 } 11689 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom(java.io.InputStream input) 11690 throws java.io.IOException { 11691 return PARSER.parseFrom(input); 11692 } 11693 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11694 java.io.InputStream input, 11695 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11696 throws java.io.IOException { 11697 return PARSER.parseFrom(input, extensionRegistry); 11698 } 11699 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseDelimitedFrom(java.io.InputStream input) 11700 throws java.io.IOException { 11701 return PARSER.parseDelimitedFrom(input); 11702 } 11703 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseDelimitedFrom( 11704 java.io.InputStream input, 11705 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11706 throws java.io.IOException { 11707 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11708 } 11709 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11710 com.google.protobuf.CodedInputStream input) 11711 throws java.io.IOException { 11712 return PARSER.parseFrom(input); 11713 } 11714 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 11715 com.google.protobuf.CodedInputStream input, 11716 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11717 throws java.io.IOException { 11718 return PARSER.parseFrom(input, extensionRegistry); 11719 } 11720 11721 public static Builder newBuilder() { return Builder.create(); } 11722 public Builder newBuilderForType() { return newBuilder(); } 11723 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection prototype) { 11724 return newBuilder().mergeFrom(prototype); 11725 } 11726 public Builder toBuilder() { return newBuilder(this); } 11727 11728 @java.lang.Override 11729 protected Builder newBuilderForType( 11730 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11731 Builder builder = new Builder(parent); 11732 return builder; 11733 } 11734 /** 11735 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection} 11736 */ 11737 public static final class Builder extends 11738 com.google.protobuf.GeneratedMessage.Builder<Builder> 11739 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSectionOrBuilder { 11740 public static final com.google.protobuf.Descriptors.Descriptor 11741 getDescriptor() { 11742 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 11743 } 11744 11745 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11746 internalGetFieldAccessorTable() { 11747 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable 11748 .ensureFieldAccessorsInitialized( 11749 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.Builder.class); 11750 } 11751 11752 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.newBuilder() 11753 private Builder() { 11754 maybeForceBuilderInitialization(); 11755 } 11756 11757 private Builder( 11758 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11759 super(parent); 11760 maybeForceBuilderInitialization(); 11761 } 11762 private void maybeForceBuilderInitialization() { 11763 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 11764 } 11765 } 11766 private static Builder create() { 11767 return new Builder(); 11768 } 11769 11770 public Builder clear() { 11771 super.clear(); 11772 return this; 11773 } 11774 11775 public Builder clone() { 11776 return create().mergeFrom(buildPartial()); 11777 } 11778 11779 public com.google.protobuf.Descriptors.Descriptor 11780 getDescriptorForType() { 11781 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 11782 } 11783 11784 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection getDefaultInstanceForType() { 11785 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.getDefaultInstance(); 11786 } 11787 11788 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection build() { 11789 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection result = buildPartial(); 11790 if (!result.isInitialized()) { 11791 throw newUninitializedMessageException(result); 11792 } 11793 return result; 11794 } 11795 11796 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection buildPartial() { 11797 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection(this); 11798 onBuilt(); 11799 return result; 11800 } 11801 11802 public Builder mergeFrom(com.google.protobuf.Message other) { 11803 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection) { 11804 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection)other); 11805 } else { 11806 super.mergeFrom(other); 11807 return this; 11808 } 11809 } 11810 11811 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection other) { 11812 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.getDefaultInstance()) return this; 11813 this.mergeUnknownFields(other.getUnknownFields()); 11814 return this; 11815 } 11816 11817 public final boolean isInitialized() { 11818 return true; 11819 } 11820 11821 public Builder mergeFrom( 11822 com.google.protobuf.CodedInputStream input, 11823 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11824 throws java.io.IOException { 11825 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parsedMessage = null; 11826 try { 11827 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 11828 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11829 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection) e.getUnfinishedMessage(); 11830 throw e; 11831 } finally { 11832 if (parsedMessage != null) { 11833 mergeFrom(parsedMessage); 11834 } 11835 } 11836 return this; 11837 } 11838 11839 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeReferenceSection) 11840 } 11841 11842 static { 11843 defaultInstance = new INodeReferenceSection(true); 11844 defaultInstance.initFields(); 11845 } 11846 11847 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeReferenceSection) 11848 } 11849 11850 public interface SnapshotSectionOrBuilder 11851 extends com.google.protobuf.MessageOrBuilder { 11852 11853 // optional uint32 snapshotCounter = 1; 11854 /** 11855 * <code>optional uint32 snapshotCounter = 1;</code> 11856 */ 11857 boolean hasSnapshotCounter(); 11858 /** 11859 * <code>optional uint32 snapshotCounter = 1;</code> 11860 */ 11861 int getSnapshotCounter(); 11862 11863 // repeated uint64 snapshottableDir = 2 [packed = true]; 11864 /** 11865 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 11866 */ 11867 java.util.List<java.lang.Long> getSnapshottableDirList(); 11868 /** 11869 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 11870 */ 11871 int getSnapshottableDirCount(); 11872 /** 11873 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 11874 */ 11875 long getSnapshottableDir(int index); 11876 11877 // optional uint32 numSnapshots = 3; 11878 /** 11879 * <code>optional uint32 numSnapshots = 3;</code> 11880 * 11881 * <pre> 11882 * total number of snapshots 11883 * </pre> 11884 */ 11885 boolean hasNumSnapshots(); 11886 /** 11887 * <code>optional uint32 numSnapshots = 3;</code> 11888 * 11889 * <pre> 11890 * total number of snapshots 11891 * </pre> 11892 */ 11893 int getNumSnapshots(); 11894 } 11895 /** 11896 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection} 11897 * 11898 * <pre> 11899 ** 11900 * This section records the information about snapshot 11901 * NAME: SNAPSHOT 11902 * </pre> 11903 */ 11904 public static final class SnapshotSection extends 11905 com.google.protobuf.GeneratedMessage 11906 implements SnapshotSectionOrBuilder { 11907 // Use SnapshotSection.newBuilder() to construct. 11908 private SnapshotSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 11909 super(builder); 11910 this.unknownFields = builder.getUnknownFields(); 11911 } 11912 private SnapshotSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 11913 11914 private static final SnapshotSection defaultInstance; 11915 public static SnapshotSection getDefaultInstance() { 11916 return defaultInstance; 11917 } 11918 11919 public SnapshotSection getDefaultInstanceForType() { 11920 return defaultInstance; 11921 } 11922 11923 private final com.google.protobuf.UnknownFieldSet unknownFields; 11924 @java.lang.Override 11925 public final com.google.protobuf.UnknownFieldSet 11926 getUnknownFields() { 11927 return this.unknownFields; 11928 } 11929 private SnapshotSection( 11930 com.google.protobuf.CodedInputStream input, 11931 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11932 throws com.google.protobuf.InvalidProtocolBufferException { 11933 initFields(); 11934 int mutable_bitField0_ = 0; 11935 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 11936 com.google.protobuf.UnknownFieldSet.newBuilder(); 11937 try { 11938 boolean done = false; 11939 while (!done) { 11940 int tag = input.readTag(); 11941 switch (tag) { 11942 case 0: 11943 done = true; 11944 break; 11945 default: { 11946 if (!parseUnknownField(input, unknownFields, 11947 extensionRegistry, tag)) { 11948 done = true; 11949 } 11950 break; 11951 } 11952 case 8: { 11953 bitField0_ |= 0x00000001; 11954 snapshotCounter_ = input.readUInt32(); 11955 break; 11956 } 11957 case 16: { 11958 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 11959 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(); 11960 mutable_bitField0_ |= 0x00000002; 11961 } 11962 snapshottableDir_.add(input.readUInt64()); 11963 break; 11964 } 11965 case 18: { 11966 int length = input.readRawVarint32(); 11967 int limit = input.pushLimit(length); 11968 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) { 11969 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(); 11970 mutable_bitField0_ |= 0x00000002; 11971 } 11972 while (input.getBytesUntilLimit() > 0) { 11973 snapshottableDir_.add(input.readUInt64()); 11974 } 11975 input.popLimit(limit); 11976 break; 11977 } 11978 case 24: { 11979 bitField0_ |= 0x00000002; 11980 numSnapshots_ = input.readUInt32(); 11981 break; 11982 } 11983 } 11984 } 11985 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11986 throw e.setUnfinishedMessage(this); 11987 } catch (java.io.IOException e) { 11988 throw new com.google.protobuf.InvalidProtocolBufferException( 11989 e.getMessage()).setUnfinishedMessage(this); 11990 } finally { 11991 if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 11992 snapshottableDir_ = java.util.Collections.unmodifiableList(snapshottableDir_); 11993 } 11994 this.unknownFields = unknownFields.build(); 11995 makeExtensionsImmutable(); 11996 } 11997 } 11998 public static final com.google.protobuf.Descriptors.Descriptor 11999 getDescriptor() { 12000 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 12001 } 12002 12003 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12004 internalGetFieldAccessorTable() { 12005 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable 12006 .ensureFieldAccessorsInitialized( 12007 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Builder.class); 12008 } 12009 12010 public static com.google.protobuf.Parser<SnapshotSection> PARSER = 12011 new com.google.protobuf.AbstractParser<SnapshotSection>() { 12012 public SnapshotSection parsePartialFrom( 12013 com.google.protobuf.CodedInputStream input, 12014 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12015 throws com.google.protobuf.InvalidProtocolBufferException { 12016 return new SnapshotSection(input, extensionRegistry); 12017 } 12018 }; 12019 12020 @java.lang.Override 12021 public com.google.protobuf.Parser<SnapshotSection> getParserForType() { 12022 return PARSER; 12023 } 12024 12025 public interface SnapshotOrBuilder 12026 extends com.google.protobuf.MessageOrBuilder { 12027 12028 // optional uint32 snapshotId = 1; 12029 /** 12030 * <code>optional uint32 snapshotId = 1;</code> 12031 */ 12032 boolean hasSnapshotId(); 12033 /** 12034 * <code>optional uint32 snapshotId = 1;</code> 12035 */ 12036 int getSnapshotId(); 12037 12038 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 12039 /** 12040 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12041 * 12042 * <pre> 12043 * Snapshot root 12044 * </pre> 12045 */ 12046 boolean hasRoot(); 12047 /** 12048 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12049 * 12050 * <pre> 12051 * Snapshot root 12052 * </pre> 12053 */ 12054 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot(); 12055 /** 12056 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12057 * 12058 * <pre> 12059 * Snapshot root 12060 * </pre> 12061 */ 12062 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder(); 12063 } 12064 /** 12065 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection.Snapshot} 12066 */ 12067 public static final class Snapshot extends 12068 com.google.protobuf.GeneratedMessage 12069 implements SnapshotOrBuilder { 12070 // Use Snapshot.newBuilder() to construct. 12071 private Snapshot(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 12072 super(builder); 12073 this.unknownFields = builder.getUnknownFields(); 12074 } 12075 private Snapshot(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 12076 12077 private static final Snapshot defaultInstance; 12078 public static Snapshot getDefaultInstance() { 12079 return defaultInstance; 12080 } 12081 12082 public Snapshot getDefaultInstanceForType() { 12083 return defaultInstance; 12084 } 12085 12086 private final com.google.protobuf.UnknownFieldSet unknownFields; 12087 @java.lang.Override 12088 public final com.google.protobuf.UnknownFieldSet 12089 getUnknownFields() { 12090 return this.unknownFields; 12091 } 12092 private Snapshot( 12093 com.google.protobuf.CodedInputStream input, 12094 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12095 throws com.google.protobuf.InvalidProtocolBufferException { 12096 initFields(); 12097 int mutable_bitField0_ = 0; 12098 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 12099 com.google.protobuf.UnknownFieldSet.newBuilder(); 12100 try { 12101 boolean done = false; 12102 while (!done) { 12103 int tag = input.readTag(); 12104 switch (tag) { 12105 case 0: 12106 done = true; 12107 break; 12108 default: { 12109 if (!parseUnknownField(input, unknownFields, 12110 extensionRegistry, tag)) { 12111 done = true; 12112 } 12113 break; 12114 } 12115 case 8: { 12116 bitField0_ |= 0x00000001; 12117 snapshotId_ = input.readUInt32(); 12118 break; 12119 } 12120 case 18: { 12121 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder subBuilder = null; 12122 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12123 subBuilder = root_.toBuilder(); 12124 } 12125 root_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.PARSER, extensionRegistry); 12126 if (subBuilder != null) { 12127 subBuilder.mergeFrom(root_); 12128 root_ = subBuilder.buildPartial(); 12129 } 12130 bitField0_ |= 0x00000002; 12131 break; 12132 } 12133 } 12134 } 12135 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12136 throw e.setUnfinishedMessage(this); 12137 } catch (java.io.IOException e) { 12138 throw new com.google.protobuf.InvalidProtocolBufferException( 12139 e.getMessage()).setUnfinishedMessage(this); 12140 } finally { 12141 this.unknownFields = unknownFields.build(); 12142 makeExtensionsImmutable(); 12143 } 12144 } 12145 public static final com.google.protobuf.Descriptors.Descriptor 12146 getDescriptor() { 12147 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 12148 } 12149 12150 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12151 internalGetFieldAccessorTable() { 12152 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable 12153 .ensureFieldAccessorsInitialized( 12154 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.Builder.class); 12155 } 12156 12157 public static com.google.protobuf.Parser<Snapshot> PARSER = 12158 new com.google.protobuf.AbstractParser<Snapshot>() { 12159 public Snapshot parsePartialFrom( 12160 com.google.protobuf.CodedInputStream input, 12161 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12162 throws com.google.protobuf.InvalidProtocolBufferException { 12163 return new Snapshot(input, extensionRegistry); 12164 } 12165 }; 12166 12167 @java.lang.Override 12168 public com.google.protobuf.Parser<Snapshot> getParserForType() { 12169 return PARSER; 12170 } 12171 12172 private int bitField0_; 12173 // optional uint32 snapshotId = 1; 12174 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 12175 private int snapshotId_; 12176 /** 12177 * <code>optional uint32 snapshotId = 1;</code> 12178 */ 12179 public boolean hasSnapshotId() { 12180 return ((bitField0_ & 0x00000001) == 0x00000001); 12181 } 12182 /** 12183 * <code>optional uint32 snapshotId = 1;</code> 12184 */ 12185 public int getSnapshotId() { 12186 return snapshotId_; 12187 } 12188 12189 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 12190 public static final int ROOT_FIELD_NUMBER = 2; 12191 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode root_; 12192 /** 12193 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12194 * 12195 * <pre> 12196 * Snapshot root 12197 * </pre> 12198 */ 12199 public boolean hasRoot() { 12200 return ((bitField0_ & 0x00000002) == 0x00000002); 12201 } 12202 /** 12203 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12204 * 12205 * <pre> 12206 * Snapshot root 12207 * </pre> 12208 */ 12209 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot() { 12210 return root_; 12211 } 12212 /** 12213 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12214 * 12215 * <pre> 12216 * Snapshot root 12217 * </pre> 12218 */ 12219 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder() { 12220 return root_; 12221 } 12222 12223 private void initFields() { 12224 snapshotId_ = 0; 12225 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 12226 } 12227 private byte memoizedIsInitialized = -1; 12228 public final boolean isInitialized() { 12229 byte isInitialized = memoizedIsInitialized; 12230 if (isInitialized != -1) return isInitialized == 1; 12231 12232 if (hasRoot()) { 12233 if (!getRoot().isInitialized()) { 12234 memoizedIsInitialized = 0; 12235 return false; 12236 } 12237 } 12238 memoizedIsInitialized = 1; 12239 return true; 12240 } 12241 12242 public void writeTo(com.google.protobuf.CodedOutputStream output) 12243 throws java.io.IOException { 12244 getSerializedSize(); 12245 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12246 output.writeUInt32(1, snapshotId_); 12247 } 12248 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12249 output.writeMessage(2, root_); 12250 } 12251 getUnknownFields().writeTo(output); 12252 } 12253 12254 private int memoizedSerializedSize = -1; 12255 public int getSerializedSize() { 12256 int size = memoizedSerializedSize; 12257 if (size != -1) return size; 12258 12259 size = 0; 12260 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12261 size += com.google.protobuf.CodedOutputStream 12262 .computeUInt32Size(1, snapshotId_); 12263 } 12264 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12265 size += com.google.protobuf.CodedOutputStream 12266 .computeMessageSize(2, root_); 12267 } 12268 size += getUnknownFields().getSerializedSize(); 12269 memoizedSerializedSize = size; 12270 return size; 12271 } 12272 12273 private static final long serialVersionUID = 0L; 12274 @java.lang.Override 12275 protected java.lang.Object writeReplace() 12276 throws java.io.ObjectStreamException { 12277 return super.writeReplace(); 12278 } 12279 12280 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12281 com.google.protobuf.ByteString data) 12282 throws com.google.protobuf.InvalidProtocolBufferException { 12283 return PARSER.parseFrom(data); 12284 } 12285 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12286 com.google.protobuf.ByteString data, 12287 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12288 throws com.google.protobuf.InvalidProtocolBufferException { 12289 return PARSER.parseFrom(data, extensionRegistry); 12290 } 12291 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom(byte[] data) 12292 throws com.google.protobuf.InvalidProtocolBufferException { 12293 return PARSER.parseFrom(data); 12294 } 12295 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12296 byte[] data, 12297 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12298 throws com.google.protobuf.InvalidProtocolBufferException { 12299 return PARSER.parseFrom(data, extensionRegistry); 12300 } 12301 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom(java.io.InputStream input) 12302 throws java.io.IOException { 12303 return PARSER.parseFrom(input); 12304 } 12305 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12306 java.io.InputStream input, 12307 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12308 throws java.io.IOException { 12309 return PARSER.parseFrom(input, extensionRegistry); 12310 } 12311 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseDelimitedFrom(java.io.InputStream input) 12312 throws java.io.IOException { 12313 return PARSER.parseDelimitedFrom(input); 12314 } 12315 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseDelimitedFrom( 12316 java.io.InputStream input, 12317 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12318 throws java.io.IOException { 12319 return PARSER.parseDelimitedFrom(input, extensionRegistry); 12320 } 12321 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12322 com.google.protobuf.CodedInputStream input) 12323 throws java.io.IOException { 12324 return PARSER.parseFrom(input); 12325 } 12326 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 12327 com.google.protobuf.CodedInputStream input, 12328 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12329 throws java.io.IOException { 12330 return PARSER.parseFrom(input, extensionRegistry); 12331 } 12332 12333 public static Builder newBuilder() { return Builder.create(); } 12334 public Builder newBuilderForType() { return newBuilder(); } 12335 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot prototype) { 12336 return newBuilder().mergeFrom(prototype); 12337 } 12338 public Builder toBuilder() { return newBuilder(this); } 12339 12340 @java.lang.Override 12341 protected Builder newBuilderForType( 12342 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12343 Builder builder = new Builder(parent); 12344 return builder; 12345 } 12346 /** 12347 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection.Snapshot} 12348 */ 12349 public static final class Builder extends 12350 com.google.protobuf.GeneratedMessage.Builder<Builder> 12351 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.SnapshotOrBuilder { 12352 public static final com.google.protobuf.Descriptors.Descriptor 12353 getDescriptor() { 12354 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 12355 } 12356 12357 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12358 internalGetFieldAccessorTable() { 12359 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable 12360 .ensureFieldAccessorsInitialized( 12361 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.Builder.class); 12362 } 12363 12364 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.newBuilder() 12365 private Builder() { 12366 maybeForceBuilderInitialization(); 12367 } 12368 12369 private Builder( 12370 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12371 super(parent); 12372 maybeForceBuilderInitialization(); 12373 } 12374 private void maybeForceBuilderInitialization() { 12375 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 12376 getRootFieldBuilder(); 12377 } 12378 } 12379 private static Builder create() { 12380 return new Builder(); 12381 } 12382 12383 public Builder clear() { 12384 super.clear(); 12385 snapshotId_ = 0; 12386 bitField0_ = (bitField0_ & ~0x00000001); 12387 if (rootBuilder_ == null) { 12388 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 12389 } else { 12390 rootBuilder_.clear(); 12391 } 12392 bitField0_ = (bitField0_ & ~0x00000002); 12393 return this; 12394 } 12395 12396 public Builder clone() { 12397 return create().mergeFrom(buildPartial()); 12398 } 12399 12400 public com.google.protobuf.Descriptors.Descriptor 12401 getDescriptorForType() { 12402 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 12403 } 12404 12405 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot getDefaultInstanceForType() { 12406 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.getDefaultInstance(); 12407 } 12408 12409 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot build() { 12410 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot result = buildPartial(); 12411 if (!result.isInitialized()) { 12412 throw newUninitializedMessageException(result); 12413 } 12414 return result; 12415 } 12416 12417 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot buildPartial() { 12418 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot(this); 12419 int from_bitField0_ = bitField0_; 12420 int to_bitField0_ = 0; 12421 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 12422 to_bitField0_ |= 0x00000001; 12423 } 12424 result.snapshotId_ = snapshotId_; 12425 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 12426 to_bitField0_ |= 0x00000002; 12427 } 12428 if (rootBuilder_ == null) { 12429 result.root_ = root_; 12430 } else { 12431 result.root_ = rootBuilder_.build(); 12432 } 12433 result.bitField0_ = to_bitField0_; 12434 onBuilt(); 12435 return result; 12436 } 12437 12438 public Builder mergeFrom(com.google.protobuf.Message other) { 12439 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot) { 12440 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot)other); 12441 } else { 12442 super.mergeFrom(other); 12443 return this; 12444 } 12445 } 12446 12447 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot other) { 12448 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.getDefaultInstance()) return this; 12449 if (other.hasSnapshotId()) { 12450 setSnapshotId(other.getSnapshotId()); 12451 } 12452 if (other.hasRoot()) { 12453 mergeRoot(other.getRoot()); 12454 } 12455 this.mergeUnknownFields(other.getUnknownFields()); 12456 return this; 12457 } 12458 12459 public final boolean isInitialized() { 12460 if (hasRoot()) { 12461 if (!getRoot().isInitialized()) { 12462 12463 return false; 12464 } 12465 } 12466 return true; 12467 } 12468 12469 public Builder mergeFrom( 12470 com.google.protobuf.CodedInputStream input, 12471 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12472 throws java.io.IOException { 12473 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parsedMessage = null; 12474 try { 12475 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 12476 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12477 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot) e.getUnfinishedMessage(); 12478 throw e; 12479 } finally { 12480 if (parsedMessage != null) { 12481 mergeFrom(parsedMessage); 12482 } 12483 } 12484 return this; 12485 } 12486 private int bitField0_; 12487 12488 // optional uint32 snapshotId = 1; 12489 private int snapshotId_ ; 12490 /** 12491 * <code>optional uint32 snapshotId = 1;</code> 12492 */ 12493 public boolean hasSnapshotId() { 12494 return ((bitField0_ & 0x00000001) == 0x00000001); 12495 } 12496 /** 12497 * <code>optional uint32 snapshotId = 1;</code> 12498 */ 12499 public int getSnapshotId() { 12500 return snapshotId_; 12501 } 12502 /** 12503 * <code>optional uint32 snapshotId = 1;</code> 12504 */ 12505 public Builder setSnapshotId(int value) { 12506 bitField0_ |= 0x00000001; 12507 snapshotId_ = value; 12508 onChanged(); 12509 return this; 12510 } 12511 /** 12512 * <code>optional uint32 snapshotId = 1;</code> 12513 */ 12514 public Builder clearSnapshotId() { 12515 bitField0_ = (bitField0_ & ~0x00000001); 12516 snapshotId_ = 0; 12517 onChanged(); 12518 return this; 12519 } 12520 12521 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 12522 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 12523 private com.google.protobuf.SingleFieldBuilder< 12524 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder> rootBuilder_; 12525 /** 12526 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12527 * 12528 * <pre> 12529 * Snapshot root 12530 * </pre> 12531 */ 12532 public boolean hasRoot() { 12533 return ((bitField0_ & 0x00000002) == 0x00000002); 12534 } 12535 /** 12536 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12537 * 12538 * <pre> 12539 * Snapshot root 12540 * </pre> 12541 */ 12542 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot() { 12543 if (rootBuilder_ == null) { 12544 return root_; 12545 } else { 12546 return rootBuilder_.getMessage(); 12547 } 12548 } 12549 /** 12550 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12551 * 12552 * <pre> 12553 * Snapshot root 12554 * </pre> 12555 */ 12556 public Builder setRoot(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode value) { 12557 if (rootBuilder_ == null) { 12558 if (value == null) { 12559 throw new NullPointerException(); 12560 } 12561 root_ = value; 12562 onChanged(); 12563 } else { 12564 rootBuilder_.setMessage(value); 12565 } 12566 bitField0_ |= 0x00000002; 12567 return this; 12568 } 12569 /** 12570 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12571 * 12572 * <pre> 12573 * Snapshot root 12574 * </pre> 12575 */ 12576 public Builder setRoot( 12577 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder builderForValue) { 12578 if (rootBuilder_ == null) { 12579 root_ = builderForValue.build(); 12580 onChanged(); 12581 } else { 12582 rootBuilder_.setMessage(builderForValue.build()); 12583 } 12584 bitField0_ |= 0x00000002; 12585 return this; 12586 } 12587 /** 12588 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12589 * 12590 * <pre> 12591 * Snapshot root 12592 * </pre> 12593 */ 12594 public Builder mergeRoot(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode value) { 12595 if (rootBuilder_ == null) { 12596 if (((bitField0_ & 0x00000002) == 0x00000002) && 12597 root_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance()) { 12598 root_ = 12599 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.newBuilder(root_).mergeFrom(value).buildPartial(); 12600 } else { 12601 root_ = value; 12602 } 12603 onChanged(); 12604 } else { 12605 rootBuilder_.mergeFrom(value); 12606 } 12607 bitField0_ |= 0x00000002; 12608 return this; 12609 } 12610 /** 12611 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12612 * 12613 * <pre> 12614 * Snapshot root 12615 * </pre> 12616 */ 12617 public Builder clearRoot() { 12618 if (rootBuilder_ == null) { 12619 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 12620 onChanged(); 12621 } else { 12622 rootBuilder_.clear(); 12623 } 12624 bitField0_ = (bitField0_ & ~0x00000002); 12625 return this; 12626 } 12627 /** 12628 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12629 * 12630 * <pre> 12631 * Snapshot root 12632 * </pre> 12633 */ 12634 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder getRootBuilder() { 12635 bitField0_ |= 0x00000002; 12636 onChanged(); 12637 return getRootFieldBuilder().getBuilder(); 12638 } 12639 /** 12640 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12641 * 12642 * <pre> 12643 * Snapshot root 12644 * </pre> 12645 */ 12646 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder() { 12647 if (rootBuilder_ != null) { 12648 return rootBuilder_.getMessageOrBuilder(); 12649 } else { 12650 return root_; 12651 } 12652 } 12653 /** 12654 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 12655 * 12656 * <pre> 12657 * Snapshot root 12658 * </pre> 12659 */ 12660 private com.google.protobuf.SingleFieldBuilder< 12661 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder> 12662 getRootFieldBuilder() { 12663 if (rootBuilder_ == null) { 12664 rootBuilder_ = new com.google.protobuf.SingleFieldBuilder< 12665 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder>( 12666 root_, 12667 getParentForChildren(), 12668 isClean()); 12669 root_ = null; 12670 } 12671 return rootBuilder_; 12672 } 12673 12674 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotSection.Snapshot) 12675 } 12676 12677 static { 12678 defaultInstance = new Snapshot(true); 12679 defaultInstance.initFields(); 12680 } 12681 12682 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotSection.Snapshot) 12683 } 12684 12685 private int bitField0_; 12686 // optional uint32 snapshotCounter = 1; 12687 public static final int SNAPSHOTCOUNTER_FIELD_NUMBER = 1; 12688 private int snapshotCounter_; 12689 /** 12690 * <code>optional uint32 snapshotCounter = 1;</code> 12691 */ 12692 public boolean hasSnapshotCounter() { 12693 return ((bitField0_ & 0x00000001) == 0x00000001); 12694 } 12695 /** 12696 * <code>optional uint32 snapshotCounter = 1;</code> 12697 */ 12698 public int getSnapshotCounter() { 12699 return snapshotCounter_; 12700 } 12701 12702 // repeated uint64 snapshottableDir = 2 [packed = true]; 12703 public static final int SNAPSHOTTABLEDIR_FIELD_NUMBER = 2; 12704 private java.util.List<java.lang.Long> snapshottableDir_; 12705 /** 12706 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 12707 */ 12708 public java.util.List<java.lang.Long> 12709 getSnapshottableDirList() { 12710 return snapshottableDir_; 12711 } 12712 /** 12713 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 12714 */ 12715 public int getSnapshottableDirCount() { 12716 return snapshottableDir_.size(); 12717 } 12718 /** 12719 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 12720 */ 12721 public long getSnapshottableDir(int index) { 12722 return snapshottableDir_.get(index); 12723 } 12724 private int snapshottableDirMemoizedSerializedSize = -1; 12725 12726 // optional uint32 numSnapshots = 3; 12727 public static final int NUMSNAPSHOTS_FIELD_NUMBER = 3; 12728 private int numSnapshots_; 12729 /** 12730 * <code>optional uint32 numSnapshots = 3;</code> 12731 * 12732 * <pre> 12733 * total number of snapshots 12734 * </pre> 12735 */ 12736 public boolean hasNumSnapshots() { 12737 return ((bitField0_ & 0x00000002) == 0x00000002); 12738 } 12739 /** 12740 * <code>optional uint32 numSnapshots = 3;</code> 12741 * 12742 * <pre> 12743 * total number of snapshots 12744 * </pre> 12745 */ 12746 public int getNumSnapshots() { 12747 return numSnapshots_; 12748 } 12749 12750 private void initFields() { 12751 snapshotCounter_ = 0; 12752 snapshottableDir_ = java.util.Collections.emptyList(); 12753 numSnapshots_ = 0; 12754 } 12755 private byte memoizedIsInitialized = -1; 12756 public final boolean isInitialized() { 12757 byte isInitialized = memoizedIsInitialized; 12758 if (isInitialized != -1) return isInitialized == 1; 12759 12760 memoizedIsInitialized = 1; 12761 return true; 12762 } 12763 12764 public void writeTo(com.google.protobuf.CodedOutputStream output) 12765 throws java.io.IOException { 12766 getSerializedSize(); 12767 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12768 output.writeUInt32(1, snapshotCounter_); 12769 } 12770 if (getSnapshottableDirList().size() > 0) { 12771 output.writeRawVarint32(18); 12772 output.writeRawVarint32(snapshottableDirMemoizedSerializedSize); 12773 } 12774 for (int i = 0; i < snapshottableDir_.size(); i++) { 12775 output.writeUInt64NoTag(snapshottableDir_.get(i)); 12776 } 12777 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12778 output.writeUInt32(3, numSnapshots_); 12779 } 12780 getUnknownFields().writeTo(output); 12781 } 12782 12783 private int memoizedSerializedSize = -1; 12784 public int getSerializedSize() { 12785 int size = memoizedSerializedSize; 12786 if (size != -1) return size; 12787 12788 size = 0; 12789 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12790 size += com.google.protobuf.CodedOutputStream 12791 .computeUInt32Size(1, snapshotCounter_); 12792 } 12793 { 12794 int dataSize = 0; 12795 for (int i = 0; i < snapshottableDir_.size(); i++) { 12796 dataSize += com.google.protobuf.CodedOutputStream 12797 .computeUInt64SizeNoTag(snapshottableDir_.get(i)); 12798 } 12799 size += dataSize; 12800 if (!getSnapshottableDirList().isEmpty()) { 12801 size += 1; 12802 size += com.google.protobuf.CodedOutputStream 12803 .computeInt32SizeNoTag(dataSize); 12804 } 12805 snapshottableDirMemoizedSerializedSize = dataSize; 12806 } 12807 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12808 size += com.google.protobuf.CodedOutputStream 12809 .computeUInt32Size(3, numSnapshots_); 12810 } 12811 size += getUnknownFields().getSerializedSize(); 12812 memoizedSerializedSize = size; 12813 return size; 12814 } 12815 12816 private static final long serialVersionUID = 0L; 12817 @java.lang.Override 12818 protected java.lang.Object writeReplace() 12819 throws java.io.ObjectStreamException { 12820 return super.writeReplace(); 12821 } 12822 12823 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12824 com.google.protobuf.ByteString data) 12825 throws com.google.protobuf.InvalidProtocolBufferException { 12826 return PARSER.parseFrom(data); 12827 } 12828 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12829 com.google.protobuf.ByteString data, 12830 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12831 throws com.google.protobuf.InvalidProtocolBufferException { 12832 return PARSER.parseFrom(data, extensionRegistry); 12833 } 12834 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom(byte[] data) 12835 throws com.google.protobuf.InvalidProtocolBufferException { 12836 return PARSER.parseFrom(data); 12837 } 12838 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12839 byte[] data, 12840 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12841 throws com.google.protobuf.InvalidProtocolBufferException { 12842 return PARSER.parseFrom(data, extensionRegistry); 12843 } 12844 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom(java.io.InputStream input) 12845 throws java.io.IOException { 12846 return PARSER.parseFrom(input); 12847 } 12848 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12849 java.io.InputStream input, 12850 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12851 throws java.io.IOException { 12852 return PARSER.parseFrom(input, extensionRegistry); 12853 } 12854 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseDelimitedFrom(java.io.InputStream input) 12855 throws java.io.IOException { 12856 return PARSER.parseDelimitedFrom(input); 12857 } 12858 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseDelimitedFrom( 12859 java.io.InputStream input, 12860 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12861 throws java.io.IOException { 12862 return PARSER.parseDelimitedFrom(input, extensionRegistry); 12863 } 12864 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12865 com.google.protobuf.CodedInputStream input) 12866 throws java.io.IOException { 12867 return PARSER.parseFrom(input); 12868 } 12869 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 12870 com.google.protobuf.CodedInputStream input, 12871 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12872 throws java.io.IOException { 12873 return PARSER.parseFrom(input, extensionRegistry); 12874 } 12875 12876 public static Builder newBuilder() { return Builder.create(); } 12877 public Builder newBuilderForType() { return newBuilder(); } 12878 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection prototype) { 12879 return newBuilder().mergeFrom(prototype); 12880 } 12881 public Builder toBuilder() { return newBuilder(this); } 12882 12883 @java.lang.Override 12884 protected Builder newBuilderForType( 12885 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12886 Builder builder = new Builder(parent); 12887 return builder; 12888 } 12889 /** 12890 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection} 12891 * 12892 * <pre> 12893 ** 12894 * This section records the information about snapshot 12895 * NAME: SNAPSHOT 12896 * </pre> 12897 */ 12898 public static final class Builder extends 12899 com.google.protobuf.GeneratedMessage.Builder<Builder> 12900 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSectionOrBuilder { 12901 public static final com.google.protobuf.Descriptors.Descriptor 12902 getDescriptor() { 12903 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 12904 } 12905 12906 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12907 internalGetFieldAccessorTable() { 12908 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable 12909 .ensureFieldAccessorsInitialized( 12910 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Builder.class); 12911 } 12912 12913 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.newBuilder() 12914 private Builder() { 12915 maybeForceBuilderInitialization(); 12916 } 12917 12918 private Builder( 12919 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12920 super(parent); 12921 maybeForceBuilderInitialization(); 12922 } 12923 private void maybeForceBuilderInitialization() { 12924 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 12925 } 12926 } 12927 private static Builder create() { 12928 return new Builder(); 12929 } 12930 12931 public Builder clear() { 12932 super.clear(); 12933 snapshotCounter_ = 0; 12934 bitField0_ = (bitField0_ & ~0x00000001); 12935 snapshottableDir_ = java.util.Collections.emptyList(); 12936 bitField0_ = (bitField0_ & ~0x00000002); 12937 numSnapshots_ = 0; 12938 bitField0_ = (bitField0_ & ~0x00000004); 12939 return this; 12940 } 12941 12942 public Builder clone() { 12943 return create().mergeFrom(buildPartial()); 12944 } 12945 12946 public com.google.protobuf.Descriptors.Descriptor 12947 getDescriptorForType() { 12948 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 12949 } 12950 12951 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection getDefaultInstanceForType() { 12952 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.getDefaultInstance(); 12953 } 12954 12955 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection build() { 12956 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection result = buildPartial(); 12957 if (!result.isInitialized()) { 12958 throw newUninitializedMessageException(result); 12959 } 12960 return result; 12961 } 12962 12963 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection buildPartial() { 12964 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection(this); 12965 int from_bitField0_ = bitField0_; 12966 int to_bitField0_ = 0; 12967 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 12968 to_bitField0_ |= 0x00000001; 12969 } 12970 result.snapshotCounter_ = snapshotCounter_; 12971 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12972 snapshottableDir_ = java.util.Collections.unmodifiableList(snapshottableDir_); 12973 bitField0_ = (bitField0_ & ~0x00000002); 12974 } 12975 result.snapshottableDir_ = snapshottableDir_; 12976 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 12977 to_bitField0_ |= 0x00000002; 12978 } 12979 result.numSnapshots_ = numSnapshots_; 12980 result.bitField0_ = to_bitField0_; 12981 onBuilt(); 12982 return result; 12983 } 12984 12985 public Builder mergeFrom(com.google.protobuf.Message other) { 12986 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection) { 12987 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection)other); 12988 } else { 12989 super.mergeFrom(other); 12990 return this; 12991 } 12992 } 12993 12994 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection other) { 12995 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.getDefaultInstance()) return this; 12996 if (other.hasSnapshotCounter()) { 12997 setSnapshotCounter(other.getSnapshotCounter()); 12998 } 12999 if (!other.snapshottableDir_.isEmpty()) { 13000 if (snapshottableDir_.isEmpty()) { 13001 snapshottableDir_ = other.snapshottableDir_; 13002 bitField0_ = (bitField0_ & ~0x00000002); 13003 } else { 13004 ensureSnapshottableDirIsMutable(); 13005 snapshottableDir_.addAll(other.snapshottableDir_); 13006 } 13007 onChanged(); 13008 } 13009 if (other.hasNumSnapshots()) { 13010 setNumSnapshots(other.getNumSnapshots()); 13011 } 13012 this.mergeUnknownFields(other.getUnknownFields()); 13013 return this; 13014 } 13015 13016 public final boolean isInitialized() { 13017 return true; 13018 } 13019 13020 public Builder mergeFrom( 13021 com.google.protobuf.CodedInputStream input, 13022 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13023 throws java.io.IOException { 13024 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parsedMessage = null; 13025 try { 13026 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 13027 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13028 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection) e.getUnfinishedMessage(); 13029 throw e; 13030 } finally { 13031 if (parsedMessage != null) { 13032 mergeFrom(parsedMessage); 13033 } 13034 } 13035 return this; 13036 } 13037 private int bitField0_; 13038 13039 // optional uint32 snapshotCounter = 1; 13040 private int snapshotCounter_ ; 13041 /** 13042 * <code>optional uint32 snapshotCounter = 1;</code> 13043 */ 13044 public boolean hasSnapshotCounter() { 13045 return ((bitField0_ & 0x00000001) == 0x00000001); 13046 } 13047 /** 13048 * <code>optional uint32 snapshotCounter = 1;</code> 13049 */ 13050 public int getSnapshotCounter() { 13051 return snapshotCounter_; 13052 } 13053 /** 13054 * <code>optional uint32 snapshotCounter = 1;</code> 13055 */ 13056 public Builder setSnapshotCounter(int value) { 13057 bitField0_ |= 0x00000001; 13058 snapshotCounter_ = value; 13059 onChanged(); 13060 return this; 13061 } 13062 /** 13063 * <code>optional uint32 snapshotCounter = 1;</code> 13064 */ 13065 public Builder clearSnapshotCounter() { 13066 bitField0_ = (bitField0_ & ~0x00000001); 13067 snapshotCounter_ = 0; 13068 onChanged(); 13069 return this; 13070 } 13071 13072 // repeated uint64 snapshottableDir = 2 [packed = true]; 13073 private java.util.List<java.lang.Long> snapshottableDir_ = java.util.Collections.emptyList(); 13074 private void ensureSnapshottableDirIsMutable() { 13075 if (!((bitField0_ & 0x00000002) == 0x00000002)) { 13076 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(snapshottableDir_); 13077 bitField0_ |= 0x00000002; 13078 } 13079 } 13080 /** 13081 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13082 */ 13083 public java.util.List<java.lang.Long> 13084 getSnapshottableDirList() { 13085 return java.util.Collections.unmodifiableList(snapshottableDir_); 13086 } 13087 /** 13088 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13089 */ 13090 public int getSnapshottableDirCount() { 13091 return snapshottableDir_.size(); 13092 } 13093 /** 13094 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13095 */ 13096 public long getSnapshottableDir(int index) { 13097 return snapshottableDir_.get(index); 13098 } 13099 /** 13100 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13101 */ 13102 public Builder setSnapshottableDir( 13103 int index, long value) { 13104 ensureSnapshottableDirIsMutable(); 13105 snapshottableDir_.set(index, value); 13106 onChanged(); 13107 return this; 13108 } 13109 /** 13110 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13111 */ 13112 public Builder addSnapshottableDir(long value) { 13113 ensureSnapshottableDirIsMutable(); 13114 snapshottableDir_.add(value); 13115 onChanged(); 13116 return this; 13117 } 13118 /** 13119 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13120 */ 13121 public Builder addAllSnapshottableDir( 13122 java.lang.Iterable<? extends java.lang.Long> values) { 13123 ensureSnapshottableDirIsMutable(); 13124 super.addAll(values, snapshottableDir_); 13125 onChanged(); 13126 return this; 13127 } 13128 /** 13129 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13130 */ 13131 public Builder clearSnapshottableDir() { 13132 snapshottableDir_ = java.util.Collections.emptyList(); 13133 bitField0_ = (bitField0_ & ~0x00000002); 13134 onChanged(); 13135 return this; 13136 } 13137 13138 // optional uint32 numSnapshots = 3; 13139 private int numSnapshots_ ; 13140 /** 13141 * <code>optional uint32 numSnapshots = 3;</code> 13142 * 13143 * <pre> 13144 * total number of snapshots 13145 * </pre> 13146 */ 13147 public boolean hasNumSnapshots() { 13148 return ((bitField0_ & 0x00000004) == 0x00000004); 13149 } 13150 /** 13151 * <code>optional uint32 numSnapshots = 3;</code> 13152 * 13153 * <pre> 13154 * total number of snapshots 13155 * </pre> 13156 */ 13157 public int getNumSnapshots() { 13158 return numSnapshots_; 13159 } 13160 /** 13161 * <code>optional uint32 numSnapshots = 3;</code> 13162 * 13163 * <pre> 13164 * total number of snapshots 13165 * </pre> 13166 */ 13167 public Builder setNumSnapshots(int value) { 13168 bitField0_ |= 0x00000004; 13169 numSnapshots_ = value; 13170 onChanged(); 13171 return this; 13172 } 13173 /** 13174 * <code>optional uint32 numSnapshots = 3;</code> 13175 * 13176 * <pre> 13177 * total number of snapshots 13178 * </pre> 13179 */ 13180 public Builder clearNumSnapshots() { 13181 bitField0_ = (bitField0_ & ~0x00000004); 13182 numSnapshots_ = 0; 13183 onChanged(); 13184 return this; 13185 } 13186 13187 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotSection) 13188 } 13189 13190 static { 13191 defaultInstance = new SnapshotSection(true); 13192 defaultInstance.initFields(); 13193 } 13194 13195 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotSection) 13196 } 13197 13198 public interface SnapshotDiffSectionOrBuilder 13199 extends com.google.protobuf.MessageOrBuilder { 13200 } 13201 /** 13202 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection} 13203 * 13204 * <pre> 13205 ** 13206 * This section records information about snapshot diffs 13207 * NAME: SNAPSHOT_DIFF 13208 * </pre> 13209 */ 13210 public static final class SnapshotDiffSection extends 13211 com.google.protobuf.GeneratedMessage 13212 implements SnapshotDiffSectionOrBuilder { 13213 // Use SnapshotDiffSection.newBuilder() to construct. 13214 private SnapshotDiffSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13215 super(builder); 13216 this.unknownFields = builder.getUnknownFields(); 13217 } 13218 private SnapshotDiffSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13219 13220 private static final SnapshotDiffSection defaultInstance; 13221 public static SnapshotDiffSection getDefaultInstance() { 13222 return defaultInstance; 13223 } 13224 13225 public SnapshotDiffSection getDefaultInstanceForType() { 13226 return defaultInstance; 13227 } 13228 13229 private final com.google.protobuf.UnknownFieldSet unknownFields; 13230 @java.lang.Override 13231 public final com.google.protobuf.UnknownFieldSet 13232 getUnknownFields() { 13233 return this.unknownFields; 13234 } 13235 private SnapshotDiffSection( 13236 com.google.protobuf.CodedInputStream input, 13237 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13238 throws com.google.protobuf.InvalidProtocolBufferException { 13239 initFields(); 13240 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13241 com.google.protobuf.UnknownFieldSet.newBuilder(); 13242 try { 13243 boolean done = false; 13244 while (!done) { 13245 int tag = input.readTag(); 13246 switch (tag) { 13247 case 0: 13248 done = true; 13249 break; 13250 default: { 13251 if (!parseUnknownField(input, unknownFields, 13252 extensionRegistry, tag)) { 13253 done = true; 13254 } 13255 break; 13256 } 13257 } 13258 } 13259 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13260 throw e.setUnfinishedMessage(this); 13261 } catch (java.io.IOException e) { 13262 throw new com.google.protobuf.InvalidProtocolBufferException( 13263 e.getMessage()).setUnfinishedMessage(this); 13264 } finally { 13265 this.unknownFields = unknownFields.build(); 13266 makeExtensionsImmutable(); 13267 } 13268 } 13269 public static final com.google.protobuf.Descriptors.Descriptor 13270 getDescriptor() { 13271 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 13272 } 13273 13274 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13275 internalGetFieldAccessorTable() { 13276 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable 13277 .ensureFieldAccessorsInitialized( 13278 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.Builder.class); 13279 } 13280 13281 public static com.google.protobuf.Parser<SnapshotDiffSection> PARSER = 13282 new com.google.protobuf.AbstractParser<SnapshotDiffSection>() { 13283 public SnapshotDiffSection parsePartialFrom( 13284 com.google.protobuf.CodedInputStream input, 13285 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13286 throws com.google.protobuf.InvalidProtocolBufferException { 13287 return new SnapshotDiffSection(input, extensionRegistry); 13288 } 13289 }; 13290 13291 @java.lang.Override 13292 public com.google.protobuf.Parser<SnapshotDiffSection> getParserForType() { 13293 return PARSER; 13294 } 13295 13296 public interface CreatedListEntryOrBuilder 13297 extends com.google.protobuf.MessageOrBuilder { 13298 13299 // optional bytes name = 1; 13300 /** 13301 * <code>optional bytes name = 1;</code> 13302 */ 13303 boolean hasName(); 13304 /** 13305 * <code>optional bytes name = 1;</code> 13306 */ 13307 com.google.protobuf.ByteString getName(); 13308 } 13309 /** 13310 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry} 13311 */ 13312 public static final class CreatedListEntry extends 13313 com.google.protobuf.GeneratedMessage 13314 implements CreatedListEntryOrBuilder { 13315 // Use CreatedListEntry.newBuilder() to construct. 13316 private CreatedListEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13317 super(builder); 13318 this.unknownFields = builder.getUnknownFields(); 13319 } 13320 private CreatedListEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13321 13322 private static final CreatedListEntry defaultInstance; 13323 public static CreatedListEntry getDefaultInstance() { 13324 return defaultInstance; 13325 } 13326 13327 public CreatedListEntry getDefaultInstanceForType() { 13328 return defaultInstance; 13329 } 13330 13331 private final com.google.protobuf.UnknownFieldSet unknownFields; 13332 @java.lang.Override 13333 public final com.google.protobuf.UnknownFieldSet 13334 getUnknownFields() { 13335 return this.unknownFields; 13336 } 13337 private CreatedListEntry( 13338 com.google.protobuf.CodedInputStream input, 13339 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13340 throws com.google.protobuf.InvalidProtocolBufferException { 13341 initFields(); 13342 int mutable_bitField0_ = 0; 13343 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13344 com.google.protobuf.UnknownFieldSet.newBuilder(); 13345 try { 13346 boolean done = false; 13347 while (!done) { 13348 int tag = input.readTag(); 13349 switch (tag) { 13350 case 0: 13351 done = true; 13352 break; 13353 default: { 13354 if (!parseUnknownField(input, unknownFields, 13355 extensionRegistry, tag)) { 13356 done = true; 13357 } 13358 break; 13359 } 13360 case 10: { 13361 bitField0_ |= 0x00000001; 13362 name_ = input.readBytes(); 13363 break; 13364 } 13365 } 13366 } 13367 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13368 throw e.setUnfinishedMessage(this); 13369 } catch (java.io.IOException e) { 13370 throw new com.google.protobuf.InvalidProtocolBufferException( 13371 e.getMessage()).setUnfinishedMessage(this); 13372 } finally { 13373 this.unknownFields = unknownFields.build(); 13374 makeExtensionsImmutable(); 13375 } 13376 } 13377 public static final com.google.protobuf.Descriptors.Descriptor 13378 getDescriptor() { 13379 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 13380 } 13381 13382 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13383 internalGetFieldAccessorTable() { 13384 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable 13385 .ensureFieldAccessorsInitialized( 13386 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.Builder.class); 13387 } 13388 13389 public static com.google.protobuf.Parser<CreatedListEntry> PARSER = 13390 new com.google.protobuf.AbstractParser<CreatedListEntry>() { 13391 public CreatedListEntry parsePartialFrom( 13392 com.google.protobuf.CodedInputStream input, 13393 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13394 throws com.google.protobuf.InvalidProtocolBufferException { 13395 return new CreatedListEntry(input, extensionRegistry); 13396 } 13397 }; 13398 13399 @java.lang.Override 13400 public com.google.protobuf.Parser<CreatedListEntry> getParserForType() { 13401 return PARSER; 13402 } 13403 13404 private int bitField0_; 13405 // optional bytes name = 1; 13406 public static final int NAME_FIELD_NUMBER = 1; 13407 private com.google.protobuf.ByteString name_; 13408 /** 13409 * <code>optional bytes name = 1;</code> 13410 */ 13411 public boolean hasName() { 13412 return ((bitField0_ & 0x00000001) == 0x00000001); 13413 } 13414 /** 13415 * <code>optional bytes name = 1;</code> 13416 */ 13417 public com.google.protobuf.ByteString getName() { 13418 return name_; 13419 } 13420 13421 private void initFields() { 13422 name_ = com.google.protobuf.ByteString.EMPTY; 13423 } 13424 private byte memoizedIsInitialized = -1; 13425 public final boolean isInitialized() { 13426 byte isInitialized = memoizedIsInitialized; 13427 if (isInitialized != -1) return isInitialized == 1; 13428 13429 memoizedIsInitialized = 1; 13430 return true; 13431 } 13432 13433 public void writeTo(com.google.protobuf.CodedOutputStream output) 13434 throws java.io.IOException { 13435 getSerializedSize(); 13436 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13437 output.writeBytes(1, name_); 13438 } 13439 getUnknownFields().writeTo(output); 13440 } 13441 13442 private int memoizedSerializedSize = -1; 13443 public int getSerializedSize() { 13444 int size = memoizedSerializedSize; 13445 if (size != -1) return size; 13446 13447 size = 0; 13448 if (((bitField0_ & 0x00000001) == 0x00000001)) { 13449 size += com.google.protobuf.CodedOutputStream 13450 .computeBytesSize(1, name_); 13451 } 13452 size += getUnknownFields().getSerializedSize(); 13453 memoizedSerializedSize = size; 13454 return size; 13455 } 13456 13457 private static final long serialVersionUID = 0L; 13458 @java.lang.Override 13459 protected java.lang.Object writeReplace() 13460 throws java.io.ObjectStreamException { 13461 return super.writeReplace(); 13462 } 13463 13464 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13465 com.google.protobuf.ByteString data) 13466 throws com.google.protobuf.InvalidProtocolBufferException { 13467 return PARSER.parseFrom(data); 13468 } 13469 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13470 com.google.protobuf.ByteString data, 13471 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13472 throws com.google.protobuf.InvalidProtocolBufferException { 13473 return PARSER.parseFrom(data, extensionRegistry); 13474 } 13475 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom(byte[] data) 13476 throws com.google.protobuf.InvalidProtocolBufferException { 13477 return PARSER.parseFrom(data); 13478 } 13479 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13480 byte[] data, 13481 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13482 throws com.google.protobuf.InvalidProtocolBufferException { 13483 return PARSER.parseFrom(data, extensionRegistry); 13484 } 13485 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom(java.io.InputStream input) 13486 throws java.io.IOException { 13487 return PARSER.parseFrom(input); 13488 } 13489 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13490 java.io.InputStream input, 13491 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13492 throws java.io.IOException { 13493 return PARSER.parseFrom(input, extensionRegistry); 13494 } 13495 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseDelimitedFrom(java.io.InputStream input) 13496 throws java.io.IOException { 13497 return PARSER.parseDelimitedFrom(input); 13498 } 13499 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseDelimitedFrom( 13500 java.io.InputStream input, 13501 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13502 throws java.io.IOException { 13503 return PARSER.parseDelimitedFrom(input, extensionRegistry); 13504 } 13505 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13506 com.google.protobuf.CodedInputStream input) 13507 throws java.io.IOException { 13508 return PARSER.parseFrom(input); 13509 } 13510 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 13511 com.google.protobuf.CodedInputStream input, 13512 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13513 throws java.io.IOException { 13514 return PARSER.parseFrom(input, extensionRegistry); 13515 } 13516 13517 public static Builder newBuilder() { return Builder.create(); } 13518 public Builder newBuilderForType() { return newBuilder(); } 13519 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry prototype) { 13520 return newBuilder().mergeFrom(prototype); 13521 } 13522 public Builder toBuilder() { return newBuilder(this); } 13523 13524 @java.lang.Override 13525 protected Builder newBuilderForType( 13526 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13527 Builder builder = new Builder(parent); 13528 return builder; 13529 } 13530 /** 13531 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry} 13532 */ 13533 public static final class Builder extends 13534 com.google.protobuf.GeneratedMessage.Builder<Builder> 13535 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntryOrBuilder { 13536 public static final com.google.protobuf.Descriptors.Descriptor 13537 getDescriptor() { 13538 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 13539 } 13540 13541 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13542 internalGetFieldAccessorTable() { 13543 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable 13544 .ensureFieldAccessorsInitialized( 13545 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.Builder.class); 13546 } 13547 13548 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.newBuilder() 13549 private Builder() { 13550 maybeForceBuilderInitialization(); 13551 } 13552 13553 private Builder( 13554 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13555 super(parent); 13556 maybeForceBuilderInitialization(); 13557 } 13558 private void maybeForceBuilderInitialization() { 13559 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 13560 } 13561 } 13562 private static Builder create() { 13563 return new Builder(); 13564 } 13565 13566 public Builder clear() { 13567 super.clear(); 13568 name_ = com.google.protobuf.ByteString.EMPTY; 13569 bitField0_ = (bitField0_ & ~0x00000001); 13570 return this; 13571 } 13572 13573 public Builder clone() { 13574 return create().mergeFrom(buildPartial()); 13575 } 13576 13577 public com.google.protobuf.Descriptors.Descriptor 13578 getDescriptorForType() { 13579 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 13580 } 13581 13582 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry getDefaultInstanceForType() { 13583 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.getDefaultInstance(); 13584 } 13585 13586 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry build() { 13587 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry result = buildPartial(); 13588 if (!result.isInitialized()) { 13589 throw newUninitializedMessageException(result); 13590 } 13591 return result; 13592 } 13593 13594 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry buildPartial() { 13595 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry(this); 13596 int from_bitField0_ = bitField0_; 13597 int to_bitField0_ = 0; 13598 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 13599 to_bitField0_ |= 0x00000001; 13600 } 13601 result.name_ = name_; 13602 result.bitField0_ = to_bitField0_; 13603 onBuilt(); 13604 return result; 13605 } 13606 13607 public Builder mergeFrom(com.google.protobuf.Message other) { 13608 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry) { 13609 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry)other); 13610 } else { 13611 super.mergeFrom(other); 13612 return this; 13613 } 13614 } 13615 13616 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry other) { 13617 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.getDefaultInstance()) return this; 13618 if (other.hasName()) { 13619 setName(other.getName()); 13620 } 13621 this.mergeUnknownFields(other.getUnknownFields()); 13622 return this; 13623 } 13624 13625 public final boolean isInitialized() { 13626 return true; 13627 } 13628 13629 public Builder mergeFrom( 13630 com.google.protobuf.CodedInputStream input, 13631 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13632 throws java.io.IOException { 13633 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parsedMessage = null; 13634 try { 13635 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 13636 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13637 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry) e.getUnfinishedMessage(); 13638 throw e; 13639 } finally { 13640 if (parsedMessage != null) { 13641 mergeFrom(parsedMessage); 13642 } 13643 } 13644 return this; 13645 } 13646 private int bitField0_; 13647 13648 // optional bytes name = 1; 13649 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 13650 /** 13651 * <code>optional bytes name = 1;</code> 13652 */ 13653 public boolean hasName() { 13654 return ((bitField0_ & 0x00000001) == 0x00000001); 13655 } 13656 /** 13657 * <code>optional bytes name = 1;</code> 13658 */ 13659 public com.google.protobuf.ByteString getName() { 13660 return name_; 13661 } 13662 /** 13663 * <code>optional bytes name = 1;</code> 13664 */ 13665 public Builder setName(com.google.protobuf.ByteString value) { 13666 if (value == null) { 13667 throw new NullPointerException(); 13668 } 13669 bitField0_ |= 0x00000001; 13670 name_ = value; 13671 onChanged(); 13672 return this; 13673 } 13674 /** 13675 * <code>optional bytes name = 1;</code> 13676 */ 13677 public Builder clearName() { 13678 bitField0_ = (bitField0_ & ~0x00000001); 13679 name_ = getDefaultInstance().getName(); 13680 onChanged(); 13681 return this; 13682 } 13683 13684 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry) 13685 } 13686 13687 static { 13688 defaultInstance = new CreatedListEntry(true); 13689 defaultInstance.initFields(); 13690 } 13691 13692 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry) 13693 } 13694 13695 public interface DirectoryDiffOrBuilder 13696 extends com.google.protobuf.MessageOrBuilder { 13697 13698 // optional uint32 snapshotId = 1; 13699 /** 13700 * <code>optional uint32 snapshotId = 1;</code> 13701 */ 13702 boolean hasSnapshotId(); 13703 /** 13704 * <code>optional uint32 snapshotId = 1;</code> 13705 */ 13706 int getSnapshotId(); 13707 13708 // optional uint32 childrenSize = 2; 13709 /** 13710 * <code>optional uint32 childrenSize = 2;</code> 13711 */ 13712 boolean hasChildrenSize(); 13713 /** 13714 * <code>optional uint32 childrenSize = 2;</code> 13715 */ 13716 int getChildrenSize(); 13717 13718 // optional bool isSnapshotRoot = 3; 13719 /** 13720 * <code>optional bool isSnapshotRoot = 3;</code> 13721 */ 13722 boolean hasIsSnapshotRoot(); 13723 /** 13724 * <code>optional bool isSnapshotRoot = 3;</code> 13725 */ 13726 boolean getIsSnapshotRoot(); 13727 13728 // optional bytes name = 4; 13729 /** 13730 * <code>optional bytes name = 4;</code> 13731 */ 13732 boolean hasName(); 13733 /** 13734 * <code>optional bytes name = 4;</code> 13735 */ 13736 com.google.protobuf.ByteString getName(); 13737 13738 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 13739 /** 13740 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 13741 */ 13742 boolean hasSnapshotCopy(); 13743 /** 13744 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 13745 */ 13746 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy(); 13747 /** 13748 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 13749 */ 13750 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder(); 13751 13752 // optional uint32 createdListSize = 6; 13753 /** 13754 * <code>optional uint32 createdListSize = 6;</code> 13755 */ 13756 boolean hasCreatedListSize(); 13757 /** 13758 * <code>optional uint32 createdListSize = 6;</code> 13759 */ 13760 int getCreatedListSize(); 13761 13762 // repeated uint64 deletedINode = 7 [packed = true]; 13763 /** 13764 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 13765 * 13766 * <pre> 13767 * id of deleted inodes 13768 * </pre> 13769 */ 13770 java.util.List<java.lang.Long> getDeletedINodeList(); 13771 /** 13772 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 13773 * 13774 * <pre> 13775 * id of deleted inodes 13776 * </pre> 13777 */ 13778 int getDeletedINodeCount(); 13779 /** 13780 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 13781 * 13782 * <pre> 13783 * id of deleted inodes 13784 * </pre> 13785 */ 13786 long getDeletedINode(int index); 13787 13788 // repeated uint32 deletedINodeRef = 8 [packed = true]; 13789 /** 13790 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 13791 * 13792 * <pre> 13793 * id of reference nodes in the deleted list 13794 * </pre> 13795 */ 13796 java.util.List<java.lang.Integer> getDeletedINodeRefList(); 13797 /** 13798 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 13799 * 13800 * <pre> 13801 * id of reference nodes in the deleted list 13802 * </pre> 13803 */ 13804 int getDeletedINodeRefCount(); 13805 /** 13806 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 13807 * 13808 * <pre> 13809 * id of reference nodes in the deleted list 13810 * </pre> 13811 */ 13812 int getDeletedINodeRef(int index); 13813 } 13814 /** 13815 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff} 13816 */ 13817 public static final class DirectoryDiff extends 13818 com.google.protobuf.GeneratedMessage 13819 implements DirectoryDiffOrBuilder { 13820 // Use DirectoryDiff.newBuilder() to construct. 13821 private DirectoryDiff(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13822 super(builder); 13823 this.unknownFields = builder.getUnknownFields(); 13824 } 13825 private DirectoryDiff(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13826 13827 private static final DirectoryDiff defaultInstance; 13828 public static DirectoryDiff getDefaultInstance() { 13829 return defaultInstance; 13830 } 13831 13832 public DirectoryDiff getDefaultInstanceForType() { 13833 return defaultInstance; 13834 } 13835 13836 private final com.google.protobuf.UnknownFieldSet unknownFields; 13837 @java.lang.Override 13838 public final com.google.protobuf.UnknownFieldSet 13839 getUnknownFields() { 13840 return this.unknownFields; 13841 } 13842 private DirectoryDiff( 13843 com.google.protobuf.CodedInputStream input, 13844 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13845 throws com.google.protobuf.InvalidProtocolBufferException { 13846 initFields(); 13847 int mutable_bitField0_ = 0; 13848 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13849 com.google.protobuf.UnknownFieldSet.newBuilder(); 13850 try { 13851 boolean done = false; 13852 while (!done) { 13853 int tag = input.readTag(); 13854 switch (tag) { 13855 case 0: 13856 done = true; 13857 break; 13858 default: { 13859 if (!parseUnknownField(input, unknownFields, 13860 extensionRegistry, tag)) { 13861 done = true; 13862 } 13863 break; 13864 } 13865 case 8: { 13866 bitField0_ |= 0x00000001; 13867 snapshotId_ = input.readUInt32(); 13868 break; 13869 } 13870 case 16: { 13871 bitField0_ |= 0x00000002; 13872 childrenSize_ = input.readUInt32(); 13873 break; 13874 } 13875 case 24: { 13876 bitField0_ |= 0x00000004; 13877 isSnapshotRoot_ = input.readBool(); 13878 break; 13879 } 13880 case 34: { 13881 bitField0_ |= 0x00000008; 13882 name_ = input.readBytes(); 13883 break; 13884 } 13885 case 42: { 13886 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder subBuilder = null; 13887 if (((bitField0_ & 0x00000010) == 0x00000010)) { 13888 subBuilder = snapshotCopy_.toBuilder(); 13889 } 13890 snapshotCopy_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.PARSER, extensionRegistry); 13891 if (subBuilder != null) { 13892 subBuilder.mergeFrom(snapshotCopy_); 13893 snapshotCopy_ = subBuilder.buildPartial(); 13894 } 13895 bitField0_ |= 0x00000010; 13896 break; 13897 } 13898 case 48: { 13899 bitField0_ |= 0x00000020; 13900 createdListSize_ = input.readUInt32(); 13901 break; 13902 } 13903 case 56: { 13904 if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) { 13905 deletedINode_ = new java.util.ArrayList<java.lang.Long>(); 13906 mutable_bitField0_ |= 0x00000040; 13907 } 13908 deletedINode_.add(input.readUInt64()); 13909 break; 13910 } 13911 case 58: { 13912 int length = input.readRawVarint32(); 13913 int limit = input.pushLimit(length); 13914 if (!((mutable_bitField0_ & 0x00000040) == 0x00000040) && input.getBytesUntilLimit() > 0) { 13915 deletedINode_ = new java.util.ArrayList<java.lang.Long>(); 13916 mutable_bitField0_ |= 0x00000040; 13917 } 13918 while (input.getBytesUntilLimit() > 0) { 13919 deletedINode_.add(input.readUInt64()); 13920 } 13921 input.popLimit(limit); 13922 break; 13923 } 13924 case 64: { 13925 if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) { 13926 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(); 13927 mutable_bitField0_ |= 0x00000080; 13928 } 13929 deletedINodeRef_.add(input.readUInt32()); 13930 break; 13931 } 13932 case 66: { 13933 int length = input.readRawVarint32(); 13934 int limit = input.pushLimit(length); 13935 if (!((mutable_bitField0_ & 0x00000080) == 0x00000080) && input.getBytesUntilLimit() > 0) { 13936 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(); 13937 mutable_bitField0_ |= 0x00000080; 13938 } 13939 while (input.getBytesUntilLimit() > 0) { 13940 deletedINodeRef_.add(input.readUInt32()); 13941 } 13942 input.popLimit(limit); 13943 break; 13944 } 13945 } 13946 } 13947 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13948 throw e.setUnfinishedMessage(this); 13949 } catch (java.io.IOException e) { 13950 throw new com.google.protobuf.InvalidProtocolBufferException( 13951 e.getMessage()).setUnfinishedMessage(this); 13952 } finally { 13953 if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) { 13954 deletedINode_ = java.util.Collections.unmodifiableList(deletedINode_); 13955 } 13956 if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) { 13957 deletedINodeRef_ = java.util.Collections.unmodifiableList(deletedINodeRef_); 13958 } 13959 this.unknownFields = unknownFields.build(); 13960 makeExtensionsImmutable(); 13961 } 13962 } 13963 public static final com.google.protobuf.Descriptors.Descriptor 13964 getDescriptor() { 13965 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 13966 } 13967 13968 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13969 internalGetFieldAccessorTable() { 13970 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable 13971 .ensureFieldAccessorsInitialized( 13972 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.Builder.class); 13973 } 13974 13975 public static com.google.protobuf.Parser<DirectoryDiff> PARSER = 13976 new com.google.protobuf.AbstractParser<DirectoryDiff>() { 13977 public DirectoryDiff parsePartialFrom( 13978 com.google.protobuf.CodedInputStream input, 13979 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13980 throws com.google.protobuf.InvalidProtocolBufferException { 13981 return new DirectoryDiff(input, extensionRegistry); 13982 } 13983 }; 13984 13985 @java.lang.Override 13986 public com.google.protobuf.Parser<DirectoryDiff> getParserForType() { 13987 return PARSER; 13988 } 13989 13990 private int bitField0_; 13991 // optional uint32 snapshotId = 1; 13992 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 13993 private int snapshotId_; 13994 /** 13995 * <code>optional uint32 snapshotId = 1;</code> 13996 */ 13997 public boolean hasSnapshotId() { 13998 return ((bitField0_ & 0x00000001) == 0x00000001); 13999 } 14000 /** 14001 * <code>optional uint32 snapshotId = 1;</code> 14002 */ 14003 public int getSnapshotId() { 14004 return snapshotId_; 14005 } 14006 14007 // optional uint32 childrenSize = 2; 14008 public static final int CHILDRENSIZE_FIELD_NUMBER = 2; 14009 private int childrenSize_; 14010 /** 14011 * <code>optional uint32 childrenSize = 2;</code> 14012 */ 14013 public boolean hasChildrenSize() { 14014 return ((bitField0_ & 0x00000002) == 0x00000002); 14015 } 14016 /** 14017 * <code>optional uint32 childrenSize = 2;</code> 14018 */ 14019 public int getChildrenSize() { 14020 return childrenSize_; 14021 } 14022 14023 // optional bool isSnapshotRoot = 3; 14024 public static final int ISSNAPSHOTROOT_FIELD_NUMBER = 3; 14025 private boolean isSnapshotRoot_; 14026 /** 14027 * <code>optional bool isSnapshotRoot = 3;</code> 14028 */ 14029 public boolean hasIsSnapshotRoot() { 14030 return ((bitField0_ & 0x00000004) == 0x00000004); 14031 } 14032 /** 14033 * <code>optional bool isSnapshotRoot = 3;</code> 14034 */ 14035 public boolean getIsSnapshotRoot() { 14036 return isSnapshotRoot_; 14037 } 14038 14039 // optional bytes name = 4; 14040 public static final int NAME_FIELD_NUMBER = 4; 14041 private com.google.protobuf.ByteString name_; 14042 /** 14043 * <code>optional bytes name = 4;</code> 14044 */ 14045 public boolean hasName() { 14046 return ((bitField0_ & 0x00000008) == 0x00000008); 14047 } 14048 /** 14049 * <code>optional bytes name = 4;</code> 14050 */ 14051 public com.google.protobuf.ByteString getName() { 14052 return name_; 14053 } 14054 14055 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 14056 public static final int SNAPSHOTCOPY_FIELD_NUMBER = 5; 14057 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory snapshotCopy_; 14058 /** 14059 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14060 */ 14061 public boolean hasSnapshotCopy() { 14062 return ((bitField0_ & 0x00000010) == 0x00000010); 14063 } 14064 /** 14065 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14066 */ 14067 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy() { 14068 return snapshotCopy_; 14069 } 14070 /** 14071 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14072 */ 14073 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder() { 14074 return snapshotCopy_; 14075 } 14076 14077 // optional uint32 createdListSize = 6; 14078 public static final int CREATEDLISTSIZE_FIELD_NUMBER = 6; 14079 private int createdListSize_; 14080 /** 14081 * <code>optional uint32 createdListSize = 6;</code> 14082 */ 14083 public boolean hasCreatedListSize() { 14084 return ((bitField0_ & 0x00000020) == 0x00000020); 14085 } 14086 /** 14087 * <code>optional uint32 createdListSize = 6;</code> 14088 */ 14089 public int getCreatedListSize() { 14090 return createdListSize_; 14091 } 14092 14093 // repeated uint64 deletedINode = 7 [packed = true]; 14094 public static final int DELETEDINODE_FIELD_NUMBER = 7; 14095 private java.util.List<java.lang.Long> deletedINode_; 14096 /** 14097 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14098 * 14099 * <pre> 14100 * id of deleted inodes 14101 * </pre> 14102 */ 14103 public java.util.List<java.lang.Long> 14104 getDeletedINodeList() { 14105 return deletedINode_; 14106 } 14107 /** 14108 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14109 * 14110 * <pre> 14111 * id of deleted inodes 14112 * </pre> 14113 */ 14114 public int getDeletedINodeCount() { 14115 return deletedINode_.size(); 14116 } 14117 /** 14118 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14119 * 14120 * <pre> 14121 * id of deleted inodes 14122 * </pre> 14123 */ 14124 public long getDeletedINode(int index) { 14125 return deletedINode_.get(index); 14126 } 14127 private int deletedINodeMemoizedSerializedSize = -1; 14128 14129 // repeated uint32 deletedINodeRef = 8 [packed = true]; 14130 public static final int DELETEDINODEREF_FIELD_NUMBER = 8; 14131 private java.util.List<java.lang.Integer> deletedINodeRef_; 14132 /** 14133 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14134 * 14135 * <pre> 14136 * id of reference nodes in the deleted list 14137 * </pre> 14138 */ 14139 public java.util.List<java.lang.Integer> 14140 getDeletedINodeRefList() { 14141 return deletedINodeRef_; 14142 } 14143 /** 14144 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14145 * 14146 * <pre> 14147 * id of reference nodes in the deleted list 14148 * </pre> 14149 */ 14150 public int getDeletedINodeRefCount() { 14151 return deletedINodeRef_.size(); 14152 } 14153 /** 14154 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14155 * 14156 * <pre> 14157 * id of reference nodes in the deleted list 14158 * </pre> 14159 */ 14160 public int getDeletedINodeRef(int index) { 14161 return deletedINodeRef_.get(index); 14162 } 14163 private int deletedINodeRefMemoizedSerializedSize = -1; 14164 14165 private void initFields() { 14166 snapshotId_ = 0; 14167 childrenSize_ = 0; 14168 isSnapshotRoot_ = false; 14169 name_ = com.google.protobuf.ByteString.EMPTY; 14170 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 14171 createdListSize_ = 0; 14172 deletedINode_ = java.util.Collections.emptyList(); 14173 deletedINodeRef_ = java.util.Collections.emptyList(); 14174 } 14175 private byte memoizedIsInitialized = -1; 14176 public final boolean isInitialized() { 14177 byte isInitialized = memoizedIsInitialized; 14178 if (isInitialized != -1) return isInitialized == 1; 14179 14180 memoizedIsInitialized = 1; 14181 return true; 14182 } 14183 14184 public void writeTo(com.google.protobuf.CodedOutputStream output) 14185 throws java.io.IOException { 14186 getSerializedSize(); 14187 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14188 output.writeUInt32(1, snapshotId_); 14189 } 14190 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14191 output.writeUInt32(2, childrenSize_); 14192 } 14193 if (((bitField0_ & 0x00000004) == 0x00000004)) { 14194 output.writeBool(3, isSnapshotRoot_); 14195 } 14196 if (((bitField0_ & 0x00000008) == 0x00000008)) { 14197 output.writeBytes(4, name_); 14198 } 14199 if (((bitField0_ & 0x00000010) == 0x00000010)) { 14200 output.writeMessage(5, snapshotCopy_); 14201 } 14202 if (((bitField0_ & 0x00000020) == 0x00000020)) { 14203 output.writeUInt32(6, createdListSize_); 14204 } 14205 if (getDeletedINodeList().size() > 0) { 14206 output.writeRawVarint32(58); 14207 output.writeRawVarint32(deletedINodeMemoizedSerializedSize); 14208 } 14209 for (int i = 0; i < deletedINode_.size(); i++) { 14210 output.writeUInt64NoTag(deletedINode_.get(i)); 14211 } 14212 if (getDeletedINodeRefList().size() > 0) { 14213 output.writeRawVarint32(66); 14214 output.writeRawVarint32(deletedINodeRefMemoizedSerializedSize); 14215 } 14216 for (int i = 0; i < deletedINodeRef_.size(); i++) { 14217 output.writeUInt32NoTag(deletedINodeRef_.get(i)); 14218 } 14219 getUnknownFields().writeTo(output); 14220 } 14221 14222 private int memoizedSerializedSize = -1; 14223 public int getSerializedSize() { 14224 int size = memoizedSerializedSize; 14225 if (size != -1) return size; 14226 14227 size = 0; 14228 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14229 size += com.google.protobuf.CodedOutputStream 14230 .computeUInt32Size(1, snapshotId_); 14231 } 14232 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14233 size += com.google.protobuf.CodedOutputStream 14234 .computeUInt32Size(2, childrenSize_); 14235 } 14236 if (((bitField0_ & 0x00000004) == 0x00000004)) { 14237 size += com.google.protobuf.CodedOutputStream 14238 .computeBoolSize(3, isSnapshotRoot_); 14239 } 14240 if (((bitField0_ & 0x00000008) == 0x00000008)) { 14241 size += com.google.protobuf.CodedOutputStream 14242 .computeBytesSize(4, name_); 14243 } 14244 if (((bitField0_ & 0x00000010) == 0x00000010)) { 14245 size += com.google.protobuf.CodedOutputStream 14246 .computeMessageSize(5, snapshotCopy_); 14247 } 14248 if (((bitField0_ & 0x00000020) == 0x00000020)) { 14249 size += com.google.protobuf.CodedOutputStream 14250 .computeUInt32Size(6, createdListSize_); 14251 } 14252 { 14253 int dataSize = 0; 14254 for (int i = 0; i < deletedINode_.size(); i++) { 14255 dataSize += com.google.protobuf.CodedOutputStream 14256 .computeUInt64SizeNoTag(deletedINode_.get(i)); 14257 } 14258 size += dataSize; 14259 if (!getDeletedINodeList().isEmpty()) { 14260 size += 1; 14261 size += com.google.protobuf.CodedOutputStream 14262 .computeInt32SizeNoTag(dataSize); 14263 } 14264 deletedINodeMemoizedSerializedSize = dataSize; 14265 } 14266 { 14267 int dataSize = 0; 14268 for (int i = 0; i < deletedINodeRef_.size(); i++) { 14269 dataSize += com.google.protobuf.CodedOutputStream 14270 .computeUInt32SizeNoTag(deletedINodeRef_.get(i)); 14271 } 14272 size += dataSize; 14273 if (!getDeletedINodeRefList().isEmpty()) { 14274 size += 1; 14275 size += com.google.protobuf.CodedOutputStream 14276 .computeInt32SizeNoTag(dataSize); 14277 } 14278 deletedINodeRefMemoizedSerializedSize = dataSize; 14279 } 14280 size += getUnknownFields().getSerializedSize(); 14281 memoizedSerializedSize = size; 14282 return size; 14283 } 14284 14285 private static final long serialVersionUID = 0L; 14286 @java.lang.Override 14287 protected java.lang.Object writeReplace() 14288 throws java.io.ObjectStreamException { 14289 return super.writeReplace(); 14290 } 14291 14292 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14293 com.google.protobuf.ByteString data) 14294 throws com.google.protobuf.InvalidProtocolBufferException { 14295 return PARSER.parseFrom(data); 14296 } 14297 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14298 com.google.protobuf.ByteString data, 14299 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14300 throws com.google.protobuf.InvalidProtocolBufferException { 14301 return PARSER.parseFrom(data, extensionRegistry); 14302 } 14303 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom(byte[] data) 14304 throws com.google.protobuf.InvalidProtocolBufferException { 14305 return PARSER.parseFrom(data); 14306 } 14307 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14308 byte[] data, 14309 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14310 throws com.google.protobuf.InvalidProtocolBufferException { 14311 return PARSER.parseFrom(data, extensionRegistry); 14312 } 14313 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom(java.io.InputStream input) 14314 throws java.io.IOException { 14315 return PARSER.parseFrom(input); 14316 } 14317 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14318 java.io.InputStream input, 14319 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14320 throws java.io.IOException { 14321 return PARSER.parseFrom(input, extensionRegistry); 14322 } 14323 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseDelimitedFrom(java.io.InputStream input) 14324 throws java.io.IOException { 14325 return PARSER.parseDelimitedFrom(input); 14326 } 14327 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseDelimitedFrom( 14328 java.io.InputStream input, 14329 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14330 throws java.io.IOException { 14331 return PARSER.parseDelimitedFrom(input, extensionRegistry); 14332 } 14333 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14334 com.google.protobuf.CodedInputStream input) 14335 throws java.io.IOException { 14336 return PARSER.parseFrom(input); 14337 } 14338 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 14339 com.google.protobuf.CodedInputStream input, 14340 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14341 throws java.io.IOException { 14342 return PARSER.parseFrom(input, extensionRegistry); 14343 } 14344 14345 public static Builder newBuilder() { return Builder.create(); } 14346 public Builder newBuilderForType() { return newBuilder(); } 14347 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff prototype) { 14348 return newBuilder().mergeFrom(prototype); 14349 } 14350 public Builder toBuilder() { return newBuilder(this); } 14351 14352 @java.lang.Override 14353 protected Builder newBuilderForType( 14354 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14355 Builder builder = new Builder(parent); 14356 return builder; 14357 } 14358 /** 14359 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff} 14360 */ 14361 public static final class Builder extends 14362 com.google.protobuf.GeneratedMessage.Builder<Builder> 14363 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiffOrBuilder { 14364 public static final com.google.protobuf.Descriptors.Descriptor 14365 getDescriptor() { 14366 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 14367 } 14368 14369 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14370 internalGetFieldAccessorTable() { 14371 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable 14372 .ensureFieldAccessorsInitialized( 14373 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.Builder.class); 14374 } 14375 14376 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.newBuilder() 14377 private Builder() { 14378 maybeForceBuilderInitialization(); 14379 } 14380 14381 private Builder( 14382 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14383 super(parent); 14384 maybeForceBuilderInitialization(); 14385 } 14386 private void maybeForceBuilderInitialization() { 14387 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 14388 getSnapshotCopyFieldBuilder(); 14389 } 14390 } 14391 private static Builder create() { 14392 return new Builder(); 14393 } 14394 14395 public Builder clear() { 14396 super.clear(); 14397 snapshotId_ = 0; 14398 bitField0_ = (bitField0_ & ~0x00000001); 14399 childrenSize_ = 0; 14400 bitField0_ = (bitField0_ & ~0x00000002); 14401 isSnapshotRoot_ = false; 14402 bitField0_ = (bitField0_ & ~0x00000004); 14403 name_ = com.google.protobuf.ByteString.EMPTY; 14404 bitField0_ = (bitField0_ & ~0x00000008); 14405 if (snapshotCopyBuilder_ == null) { 14406 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 14407 } else { 14408 snapshotCopyBuilder_.clear(); 14409 } 14410 bitField0_ = (bitField0_ & ~0x00000010); 14411 createdListSize_ = 0; 14412 bitField0_ = (bitField0_ & ~0x00000020); 14413 deletedINode_ = java.util.Collections.emptyList(); 14414 bitField0_ = (bitField0_ & ~0x00000040); 14415 deletedINodeRef_ = java.util.Collections.emptyList(); 14416 bitField0_ = (bitField0_ & ~0x00000080); 14417 return this; 14418 } 14419 14420 public Builder clone() { 14421 return create().mergeFrom(buildPartial()); 14422 } 14423 14424 public com.google.protobuf.Descriptors.Descriptor 14425 getDescriptorForType() { 14426 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 14427 } 14428 14429 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff getDefaultInstanceForType() { 14430 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.getDefaultInstance(); 14431 } 14432 14433 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff build() { 14434 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff result = buildPartial(); 14435 if (!result.isInitialized()) { 14436 throw newUninitializedMessageException(result); 14437 } 14438 return result; 14439 } 14440 14441 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff buildPartial() { 14442 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff(this); 14443 int from_bitField0_ = bitField0_; 14444 int to_bitField0_ = 0; 14445 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 14446 to_bitField0_ |= 0x00000001; 14447 } 14448 result.snapshotId_ = snapshotId_; 14449 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 14450 to_bitField0_ |= 0x00000002; 14451 } 14452 result.childrenSize_ = childrenSize_; 14453 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 14454 to_bitField0_ |= 0x00000004; 14455 } 14456 result.isSnapshotRoot_ = isSnapshotRoot_; 14457 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 14458 to_bitField0_ |= 0x00000008; 14459 } 14460 result.name_ = name_; 14461 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 14462 to_bitField0_ |= 0x00000010; 14463 } 14464 if (snapshotCopyBuilder_ == null) { 14465 result.snapshotCopy_ = snapshotCopy_; 14466 } else { 14467 result.snapshotCopy_ = snapshotCopyBuilder_.build(); 14468 } 14469 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 14470 to_bitField0_ |= 0x00000020; 14471 } 14472 result.createdListSize_ = createdListSize_; 14473 if (((bitField0_ & 0x00000040) == 0x00000040)) { 14474 deletedINode_ = java.util.Collections.unmodifiableList(deletedINode_); 14475 bitField0_ = (bitField0_ & ~0x00000040); 14476 } 14477 result.deletedINode_ = deletedINode_; 14478 if (((bitField0_ & 0x00000080) == 0x00000080)) { 14479 deletedINodeRef_ = java.util.Collections.unmodifiableList(deletedINodeRef_); 14480 bitField0_ = (bitField0_ & ~0x00000080); 14481 } 14482 result.deletedINodeRef_ = deletedINodeRef_; 14483 result.bitField0_ = to_bitField0_; 14484 onBuilt(); 14485 return result; 14486 } 14487 14488 public Builder mergeFrom(com.google.protobuf.Message other) { 14489 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff) { 14490 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff)other); 14491 } else { 14492 super.mergeFrom(other); 14493 return this; 14494 } 14495 } 14496 14497 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff other) { 14498 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.getDefaultInstance()) return this; 14499 if (other.hasSnapshotId()) { 14500 setSnapshotId(other.getSnapshotId()); 14501 } 14502 if (other.hasChildrenSize()) { 14503 setChildrenSize(other.getChildrenSize()); 14504 } 14505 if (other.hasIsSnapshotRoot()) { 14506 setIsSnapshotRoot(other.getIsSnapshotRoot()); 14507 } 14508 if (other.hasName()) { 14509 setName(other.getName()); 14510 } 14511 if (other.hasSnapshotCopy()) { 14512 mergeSnapshotCopy(other.getSnapshotCopy()); 14513 } 14514 if (other.hasCreatedListSize()) { 14515 setCreatedListSize(other.getCreatedListSize()); 14516 } 14517 if (!other.deletedINode_.isEmpty()) { 14518 if (deletedINode_.isEmpty()) { 14519 deletedINode_ = other.deletedINode_; 14520 bitField0_ = (bitField0_ & ~0x00000040); 14521 } else { 14522 ensureDeletedINodeIsMutable(); 14523 deletedINode_.addAll(other.deletedINode_); 14524 } 14525 onChanged(); 14526 } 14527 if (!other.deletedINodeRef_.isEmpty()) { 14528 if (deletedINodeRef_.isEmpty()) { 14529 deletedINodeRef_ = other.deletedINodeRef_; 14530 bitField0_ = (bitField0_ & ~0x00000080); 14531 } else { 14532 ensureDeletedINodeRefIsMutable(); 14533 deletedINodeRef_.addAll(other.deletedINodeRef_); 14534 } 14535 onChanged(); 14536 } 14537 this.mergeUnknownFields(other.getUnknownFields()); 14538 return this; 14539 } 14540 14541 public final boolean isInitialized() { 14542 return true; 14543 } 14544 14545 public Builder mergeFrom( 14546 com.google.protobuf.CodedInputStream input, 14547 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14548 throws java.io.IOException { 14549 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parsedMessage = null; 14550 try { 14551 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 14552 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 14553 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff) e.getUnfinishedMessage(); 14554 throw e; 14555 } finally { 14556 if (parsedMessage != null) { 14557 mergeFrom(parsedMessage); 14558 } 14559 } 14560 return this; 14561 } 14562 private int bitField0_; 14563 14564 // optional uint32 snapshotId = 1; 14565 private int snapshotId_ ; 14566 /** 14567 * <code>optional uint32 snapshotId = 1;</code> 14568 */ 14569 public boolean hasSnapshotId() { 14570 return ((bitField0_ & 0x00000001) == 0x00000001); 14571 } 14572 /** 14573 * <code>optional uint32 snapshotId = 1;</code> 14574 */ 14575 public int getSnapshotId() { 14576 return snapshotId_; 14577 } 14578 /** 14579 * <code>optional uint32 snapshotId = 1;</code> 14580 */ 14581 public Builder setSnapshotId(int value) { 14582 bitField0_ |= 0x00000001; 14583 snapshotId_ = value; 14584 onChanged(); 14585 return this; 14586 } 14587 /** 14588 * <code>optional uint32 snapshotId = 1;</code> 14589 */ 14590 public Builder clearSnapshotId() { 14591 bitField0_ = (bitField0_ & ~0x00000001); 14592 snapshotId_ = 0; 14593 onChanged(); 14594 return this; 14595 } 14596 14597 // optional uint32 childrenSize = 2; 14598 private int childrenSize_ ; 14599 /** 14600 * <code>optional uint32 childrenSize = 2;</code> 14601 */ 14602 public boolean hasChildrenSize() { 14603 return ((bitField0_ & 0x00000002) == 0x00000002); 14604 } 14605 /** 14606 * <code>optional uint32 childrenSize = 2;</code> 14607 */ 14608 public int getChildrenSize() { 14609 return childrenSize_; 14610 } 14611 /** 14612 * <code>optional uint32 childrenSize = 2;</code> 14613 */ 14614 public Builder setChildrenSize(int value) { 14615 bitField0_ |= 0x00000002; 14616 childrenSize_ = value; 14617 onChanged(); 14618 return this; 14619 } 14620 /** 14621 * <code>optional uint32 childrenSize = 2;</code> 14622 */ 14623 public Builder clearChildrenSize() { 14624 bitField0_ = (bitField0_ & ~0x00000002); 14625 childrenSize_ = 0; 14626 onChanged(); 14627 return this; 14628 } 14629 14630 // optional bool isSnapshotRoot = 3; 14631 private boolean isSnapshotRoot_ ; 14632 /** 14633 * <code>optional bool isSnapshotRoot = 3;</code> 14634 */ 14635 public boolean hasIsSnapshotRoot() { 14636 return ((bitField0_ & 0x00000004) == 0x00000004); 14637 } 14638 /** 14639 * <code>optional bool isSnapshotRoot = 3;</code> 14640 */ 14641 public boolean getIsSnapshotRoot() { 14642 return isSnapshotRoot_; 14643 } 14644 /** 14645 * <code>optional bool isSnapshotRoot = 3;</code> 14646 */ 14647 public Builder setIsSnapshotRoot(boolean value) { 14648 bitField0_ |= 0x00000004; 14649 isSnapshotRoot_ = value; 14650 onChanged(); 14651 return this; 14652 } 14653 /** 14654 * <code>optional bool isSnapshotRoot = 3;</code> 14655 */ 14656 public Builder clearIsSnapshotRoot() { 14657 bitField0_ = (bitField0_ & ~0x00000004); 14658 isSnapshotRoot_ = false; 14659 onChanged(); 14660 return this; 14661 } 14662 14663 // optional bytes name = 4; 14664 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 14665 /** 14666 * <code>optional bytes name = 4;</code> 14667 */ 14668 public boolean hasName() { 14669 return ((bitField0_ & 0x00000008) == 0x00000008); 14670 } 14671 /** 14672 * <code>optional bytes name = 4;</code> 14673 */ 14674 public com.google.protobuf.ByteString getName() { 14675 return name_; 14676 } 14677 /** 14678 * <code>optional bytes name = 4;</code> 14679 */ 14680 public Builder setName(com.google.protobuf.ByteString value) { 14681 if (value == null) { 14682 throw new NullPointerException(); 14683 } 14684 bitField0_ |= 0x00000008; 14685 name_ = value; 14686 onChanged(); 14687 return this; 14688 } 14689 /** 14690 * <code>optional bytes name = 4;</code> 14691 */ 14692 public Builder clearName() { 14693 bitField0_ = (bitField0_ & ~0x00000008); 14694 name_ = getDefaultInstance().getName(); 14695 onChanged(); 14696 return this; 14697 } 14698 14699 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 14700 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 14701 private com.google.protobuf.SingleFieldBuilder< 14702 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> snapshotCopyBuilder_; 14703 /** 14704 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14705 */ 14706 public boolean hasSnapshotCopy() { 14707 return ((bitField0_ & 0x00000010) == 0x00000010); 14708 } 14709 /** 14710 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14711 */ 14712 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy() { 14713 if (snapshotCopyBuilder_ == null) { 14714 return snapshotCopy_; 14715 } else { 14716 return snapshotCopyBuilder_.getMessage(); 14717 } 14718 } 14719 /** 14720 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14721 */ 14722 public Builder setSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 14723 if (snapshotCopyBuilder_ == null) { 14724 if (value == null) { 14725 throw new NullPointerException(); 14726 } 14727 snapshotCopy_ = value; 14728 onChanged(); 14729 } else { 14730 snapshotCopyBuilder_.setMessage(value); 14731 } 14732 bitField0_ |= 0x00000010; 14733 return this; 14734 } 14735 /** 14736 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14737 */ 14738 public Builder setSnapshotCopy( 14739 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder builderForValue) { 14740 if (snapshotCopyBuilder_ == null) { 14741 snapshotCopy_ = builderForValue.build(); 14742 onChanged(); 14743 } else { 14744 snapshotCopyBuilder_.setMessage(builderForValue.build()); 14745 } 14746 bitField0_ |= 0x00000010; 14747 return this; 14748 } 14749 /** 14750 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14751 */ 14752 public Builder mergeSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 14753 if (snapshotCopyBuilder_ == null) { 14754 if (((bitField0_ & 0x00000010) == 0x00000010) && 14755 snapshotCopy_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) { 14756 snapshotCopy_ = 14757 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder(snapshotCopy_).mergeFrom(value).buildPartial(); 14758 } else { 14759 snapshotCopy_ = value; 14760 } 14761 onChanged(); 14762 } else { 14763 snapshotCopyBuilder_.mergeFrom(value); 14764 } 14765 bitField0_ |= 0x00000010; 14766 return this; 14767 } 14768 /** 14769 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14770 */ 14771 public Builder clearSnapshotCopy() { 14772 if (snapshotCopyBuilder_ == null) { 14773 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 14774 onChanged(); 14775 } else { 14776 snapshotCopyBuilder_.clear(); 14777 } 14778 bitField0_ = (bitField0_ & ~0x00000010); 14779 return this; 14780 } 14781 /** 14782 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14783 */ 14784 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder getSnapshotCopyBuilder() { 14785 bitField0_ |= 0x00000010; 14786 onChanged(); 14787 return getSnapshotCopyFieldBuilder().getBuilder(); 14788 } 14789 /** 14790 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14791 */ 14792 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder() { 14793 if (snapshotCopyBuilder_ != null) { 14794 return snapshotCopyBuilder_.getMessageOrBuilder(); 14795 } else { 14796 return snapshotCopy_; 14797 } 14798 } 14799 /** 14800 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 14801 */ 14802 private com.google.protobuf.SingleFieldBuilder< 14803 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> 14804 getSnapshotCopyFieldBuilder() { 14805 if (snapshotCopyBuilder_ == null) { 14806 snapshotCopyBuilder_ = new com.google.protobuf.SingleFieldBuilder< 14807 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder>( 14808 snapshotCopy_, 14809 getParentForChildren(), 14810 isClean()); 14811 snapshotCopy_ = null; 14812 } 14813 return snapshotCopyBuilder_; 14814 } 14815 14816 // optional uint32 createdListSize = 6; 14817 private int createdListSize_ ; 14818 /** 14819 * <code>optional uint32 createdListSize = 6;</code> 14820 */ 14821 public boolean hasCreatedListSize() { 14822 return ((bitField0_ & 0x00000020) == 0x00000020); 14823 } 14824 /** 14825 * <code>optional uint32 createdListSize = 6;</code> 14826 */ 14827 public int getCreatedListSize() { 14828 return createdListSize_; 14829 } 14830 /** 14831 * <code>optional uint32 createdListSize = 6;</code> 14832 */ 14833 public Builder setCreatedListSize(int value) { 14834 bitField0_ |= 0x00000020; 14835 createdListSize_ = value; 14836 onChanged(); 14837 return this; 14838 } 14839 /** 14840 * <code>optional uint32 createdListSize = 6;</code> 14841 */ 14842 public Builder clearCreatedListSize() { 14843 bitField0_ = (bitField0_ & ~0x00000020); 14844 createdListSize_ = 0; 14845 onChanged(); 14846 return this; 14847 } 14848 14849 // repeated uint64 deletedINode = 7 [packed = true]; 14850 private java.util.List<java.lang.Long> deletedINode_ = java.util.Collections.emptyList(); 14851 private void ensureDeletedINodeIsMutable() { 14852 if (!((bitField0_ & 0x00000040) == 0x00000040)) { 14853 deletedINode_ = new java.util.ArrayList<java.lang.Long>(deletedINode_); 14854 bitField0_ |= 0x00000040; 14855 } 14856 } 14857 /** 14858 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14859 * 14860 * <pre> 14861 * id of deleted inodes 14862 * </pre> 14863 */ 14864 public java.util.List<java.lang.Long> 14865 getDeletedINodeList() { 14866 return java.util.Collections.unmodifiableList(deletedINode_); 14867 } 14868 /** 14869 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14870 * 14871 * <pre> 14872 * id of deleted inodes 14873 * </pre> 14874 */ 14875 public int getDeletedINodeCount() { 14876 return deletedINode_.size(); 14877 } 14878 /** 14879 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14880 * 14881 * <pre> 14882 * id of deleted inodes 14883 * </pre> 14884 */ 14885 public long getDeletedINode(int index) { 14886 return deletedINode_.get(index); 14887 } 14888 /** 14889 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14890 * 14891 * <pre> 14892 * id of deleted inodes 14893 * </pre> 14894 */ 14895 public Builder setDeletedINode( 14896 int index, long value) { 14897 ensureDeletedINodeIsMutable(); 14898 deletedINode_.set(index, value); 14899 onChanged(); 14900 return this; 14901 } 14902 /** 14903 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14904 * 14905 * <pre> 14906 * id of deleted inodes 14907 * </pre> 14908 */ 14909 public Builder addDeletedINode(long value) { 14910 ensureDeletedINodeIsMutable(); 14911 deletedINode_.add(value); 14912 onChanged(); 14913 return this; 14914 } 14915 /** 14916 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14917 * 14918 * <pre> 14919 * id of deleted inodes 14920 * </pre> 14921 */ 14922 public Builder addAllDeletedINode( 14923 java.lang.Iterable<? extends java.lang.Long> values) { 14924 ensureDeletedINodeIsMutable(); 14925 super.addAll(values, deletedINode_); 14926 onChanged(); 14927 return this; 14928 } 14929 /** 14930 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 14931 * 14932 * <pre> 14933 * id of deleted inodes 14934 * </pre> 14935 */ 14936 public Builder clearDeletedINode() { 14937 deletedINode_ = java.util.Collections.emptyList(); 14938 bitField0_ = (bitField0_ & ~0x00000040); 14939 onChanged(); 14940 return this; 14941 } 14942 14943 // repeated uint32 deletedINodeRef = 8 [packed = true]; 14944 private java.util.List<java.lang.Integer> deletedINodeRef_ = java.util.Collections.emptyList(); 14945 private void ensureDeletedINodeRefIsMutable() { 14946 if (!((bitField0_ & 0x00000080) == 0x00000080)) { 14947 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(deletedINodeRef_); 14948 bitField0_ |= 0x00000080; 14949 } 14950 } 14951 /** 14952 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14953 * 14954 * <pre> 14955 * id of reference nodes in the deleted list 14956 * </pre> 14957 */ 14958 public java.util.List<java.lang.Integer> 14959 getDeletedINodeRefList() { 14960 return java.util.Collections.unmodifiableList(deletedINodeRef_); 14961 } 14962 /** 14963 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14964 * 14965 * <pre> 14966 * id of reference nodes in the deleted list 14967 * </pre> 14968 */ 14969 public int getDeletedINodeRefCount() { 14970 return deletedINodeRef_.size(); 14971 } 14972 /** 14973 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14974 * 14975 * <pre> 14976 * id of reference nodes in the deleted list 14977 * </pre> 14978 */ 14979 public int getDeletedINodeRef(int index) { 14980 return deletedINodeRef_.get(index); 14981 } 14982 /** 14983 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14984 * 14985 * <pre> 14986 * id of reference nodes in the deleted list 14987 * </pre> 14988 */ 14989 public Builder setDeletedINodeRef( 14990 int index, int value) { 14991 ensureDeletedINodeRefIsMutable(); 14992 deletedINodeRef_.set(index, value); 14993 onChanged(); 14994 return this; 14995 } 14996 /** 14997 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 14998 * 14999 * <pre> 15000 * id of reference nodes in the deleted list 15001 * </pre> 15002 */ 15003 public Builder addDeletedINodeRef(int value) { 15004 ensureDeletedINodeRefIsMutable(); 15005 deletedINodeRef_.add(value); 15006 onChanged(); 15007 return this; 15008 } 15009 /** 15010 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15011 * 15012 * <pre> 15013 * id of reference nodes in the deleted list 15014 * </pre> 15015 */ 15016 public Builder addAllDeletedINodeRef( 15017 java.lang.Iterable<? extends java.lang.Integer> values) { 15018 ensureDeletedINodeRefIsMutable(); 15019 super.addAll(values, deletedINodeRef_); 15020 onChanged(); 15021 return this; 15022 } 15023 /** 15024 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15025 * 15026 * <pre> 15027 * id of reference nodes in the deleted list 15028 * </pre> 15029 */ 15030 public Builder clearDeletedINodeRef() { 15031 deletedINodeRef_ = java.util.Collections.emptyList(); 15032 bitField0_ = (bitField0_ & ~0x00000080); 15033 onChanged(); 15034 return this; 15035 } 15036 15037 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff) 15038 } 15039 15040 static { 15041 defaultInstance = new DirectoryDiff(true); 15042 defaultInstance.initFields(); 15043 } 15044 15045 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff) 15046 } 15047 15048 public interface FileDiffOrBuilder 15049 extends com.google.protobuf.MessageOrBuilder { 15050 15051 // optional uint32 snapshotId = 1; 15052 /** 15053 * <code>optional uint32 snapshotId = 1;</code> 15054 */ 15055 boolean hasSnapshotId(); 15056 /** 15057 * <code>optional uint32 snapshotId = 1;</code> 15058 */ 15059 int getSnapshotId(); 15060 15061 // optional uint64 fileSize = 2; 15062 /** 15063 * <code>optional uint64 fileSize = 2;</code> 15064 */ 15065 boolean hasFileSize(); 15066 /** 15067 * <code>optional uint64 fileSize = 2;</code> 15068 */ 15069 long getFileSize(); 15070 15071 // optional bytes name = 3; 15072 /** 15073 * <code>optional bytes name = 3;</code> 15074 */ 15075 boolean hasName(); 15076 /** 15077 * <code>optional bytes name = 3;</code> 15078 */ 15079 com.google.protobuf.ByteString getName(); 15080 15081 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 15082 /** 15083 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15084 */ 15085 boolean hasSnapshotCopy(); 15086 /** 15087 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15088 */ 15089 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy(); 15090 /** 15091 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15092 */ 15093 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder(); 15094 } 15095 /** 15096 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff} 15097 */ 15098 public static final class FileDiff extends 15099 com.google.protobuf.GeneratedMessage 15100 implements FileDiffOrBuilder { 15101 // Use FileDiff.newBuilder() to construct. 15102 private FileDiff(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 15103 super(builder); 15104 this.unknownFields = builder.getUnknownFields(); 15105 } 15106 private FileDiff(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 15107 15108 private static final FileDiff defaultInstance; 15109 public static FileDiff getDefaultInstance() { 15110 return defaultInstance; 15111 } 15112 15113 public FileDiff getDefaultInstanceForType() { 15114 return defaultInstance; 15115 } 15116 15117 private final com.google.protobuf.UnknownFieldSet unknownFields; 15118 @java.lang.Override 15119 public final com.google.protobuf.UnknownFieldSet 15120 getUnknownFields() { 15121 return this.unknownFields; 15122 } 15123 private FileDiff( 15124 com.google.protobuf.CodedInputStream input, 15125 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15126 throws com.google.protobuf.InvalidProtocolBufferException { 15127 initFields(); 15128 int mutable_bitField0_ = 0; 15129 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15130 com.google.protobuf.UnknownFieldSet.newBuilder(); 15131 try { 15132 boolean done = false; 15133 while (!done) { 15134 int tag = input.readTag(); 15135 switch (tag) { 15136 case 0: 15137 done = true; 15138 break; 15139 default: { 15140 if (!parseUnknownField(input, unknownFields, 15141 extensionRegistry, tag)) { 15142 done = true; 15143 } 15144 break; 15145 } 15146 case 8: { 15147 bitField0_ |= 0x00000001; 15148 snapshotId_ = input.readUInt32(); 15149 break; 15150 } 15151 case 16: { 15152 bitField0_ |= 0x00000002; 15153 fileSize_ = input.readUInt64(); 15154 break; 15155 } 15156 case 26: { 15157 bitField0_ |= 0x00000004; 15158 name_ = input.readBytes(); 15159 break; 15160 } 15161 case 34: { 15162 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder subBuilder = null; 15163 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15164 subBuilder = snapshotCopy_.toBuilder(); 15165 } 15166 snapshotCopy_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.PARSER, extensionRegistry); 15167 if (subBuilder != null) { 15168 subBuilder.mergeFrom(snapshotCopy_); 15169 snapshotCopy_ = subBuilder.buildPartial(); 15170 } 15171 bitField0_ |= 0x00000008; 15172 break; 15173 } 15174 } 15175 } 15176 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15177 throw e.setUnfinishedMessage(this); 15178 } catch (java.io.IOException e) { 15179 throw new com.google.protobuf.InvalidProtocolBufferException( 15180 e.getMessage()).setUnfinishedMessage(this); 15181 } finally { 15182 this.unknownFields = unknownFields.build(); 15183 makeExtensionsImmutable(); 15184 } 15185 } 15186 public static final com.google.protobuf.Descriptors.Descriptor 15187 getDescriptor() { 15188 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 15189 } 15190 15191 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15192 internalGetFieldAccessorTable() { 15193 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable 15194 .ensureFieldAccessorsInitialized( 15195 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.Builder.class); 15196 } 15197 15198 public static com.google.protobuf.Parser<FileDiff> PARSER = 15199 new com.google.protobuf.AbstractParser<FileDiff>() { 15200 public FileDiff parsePartialFrom( 15201 com.google.protobuf.CodedInputStream input, 15202 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15203 throws com.google.protobuf.InvalidProtocolBufferException { 15204 return new FileDiff(input, extensionRegistry); 15205 } 15206 }; 15207 15208 @java.lang.Override 15209 public com.google.protobuf.Parser<FileDiff> getParserForType() { 15210 return PARSER; 15211 } 15212 15213 private int bitField0_; 15214 // optional uint32 snapshotId = 1; 15215 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 15216 private int snapshotId_; 15217 /** 15218 * <code>optional uint32 snapshotId = 1;</code> 15219 */ 15220 public boolean hasSnapshotId() { 15221 return ((bitField0_ & 0x00000001) == 0x00000001); 15222 } 15223 /** 15224 * <code>optional uint32 snapshotId = 1;</code> 15225 */ 15226 public int getSnapshotId() { 15227 return snapshotId_; 15228 } 15229 15230 // optional uint64 fileSize = 2; 15231 public static final int FILESIZE_FIELD_NUMBER = 2; 15232 private long fileSize_; 15233 /** 15234 * <code>optional uint64 fileSize = 2;</code> 15235 */ 15236 public boolean hasFileSize() { 15237 return ((bitField0_ & 0x00000002) == 0x00000002); 15238 } 15239 /** 15240 * <code>optional uint64 fileSize = 2;</code> 15241 */ 15242 public long getFileSize() { 15243 return fileSize_; 15244 } 15245 15246 // optional bytes name = 3; 15247 public static final int NAME_FIELD_NUMBER = 3; 15248 private com.google.protobuf.ByteString name_; 15249 /** 15250 * <code>optional bytes name = 3;</code> 15251 */ 15252 public boolean hasName() { 15253 return ((bitField0_ & 0x00000004) == 0x00000004); 15254 } 15255 /** 15256 * <code>optional bytes name = 3;</code> 15257 */ 15258 public com.google.protobuf.ByteString getName() { 15259 return name_; 15260 } 15261 15262 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 15263 public static final int SNAPSHOTCOPY_FIELD_NUMBER = 4; 15264 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile snapshotCopy_; 15265 /** 15266 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15267 */ 15268 public boolean hasSnapshotCopy() { 15269 return ((bitField0_ & 0x00000008) == 0x00000008); 15270 } 15271 /** 15272 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15273 */ 15274 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy() { 15275 return snapshotCopy_; 15276 } 15277 /** 15278 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15279 */ 15280 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder() { 15281 return snapshotCopy_; 15282 } 15283 15284 private void initFields() { 15285 snapshotId_ = 0; 15286 fileSize_ = 0L; 15287 name_ = com.google.protobuf.ByteString.EMPTY; 15288 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 15289 } 15290 private byte memoizedIsInitialized = -1; 15291 public final boolean isInitialized() { 15292 byte isInitialized = memoizedIsInitialized; 15293 if (isInitialized != -1) return isInitialized == 1; 15294 15295 if (hasSnapshotCopy()) { 15296 if (!getSnapshotCopy().isInitialized()) { 15297 memoizedIsInitialized = 0; 15298 return false; 15299 } 15300 } 15301 memoizedIsInitialized = 1; 15302 return true; 15303 } 15304 15305 public void writeTo(com.google.protobuf.CodedOutputStream output) 15306 throws java.io.IOException { 15307 getSerializedSize(); 15308 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15309 output.writeUInt32(1, snapshotId_); 15310 } 15311 if (((bitField0_ & 0x00000002) == 0x00000002)) { 15312 output.writeUInt64(2, fileSize_); 15313 } 15314 if (((bitField0_ & 0x00000004) == 0x00000004)) { 15315 output.writeBytes(3, name_); 15316 } 15317 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15318 output.writeMessage(4, snapshotCopy_); 15319 } 15320 getUnknownFields().writeTo(output); 15321 } 15322 15323 private int memoizedSerializedSize = -1; 15324 public int getSerializedSize() { 15325 int size = memoizedSerializedSize; 15326 if (size != -1) return size; 15327 15328 size = 0; 15329 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15330 size += com.google.protobuf.CodedOutputStream 15331 .computeUInt32Size(1, snapshotId_); 15332 } 15333 if (((bitField0_ & 0x00000002) == 0x00000002)) { 15334 size += com.google.protobuf.CodedOutputStream 15335 .computeUInt64Size(2, fileSize_); 15336 } 15337 if (((bitField0_ & 0x00000004) == 0x00000004)) { 15338 size += com.google.protobuf.CodedOutputStream 15339 .computeBytesSize(3, name_); 15340 } 15341 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15342 size += com.google.protobuf.CodedOutputStream 15343 .computeMessageSize(4, snapshotCopy_); 15344 } 15345 size += getUnknownFields().getSerializedSize(); 15346 memoizedSerializedSize = size; 15347 return size; 15348 } 15349 15350 private static final long serialVersionUID = 0L; 15351 @java.lang.Override 15352 protected java.lang.Object writeReplace() 15353 throws java.io.ObjectStreamException { 15354 return super.writeReplace(); 15355 } 15356 15357 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15358 com.google.protobuf.ByteString data) 15359 throws com.google.protobuf.InvalidProtocolBufferException { 15360 return PARSER.parseFrom(data); 15361 } 15362 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15363 com.google.protobuf.ByteString data, 15364 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15365 throws com.google.protobuf.InvalidProtocolBufferException { 15366 return PARSER.parseFrom(data, extensionRegistry); 15367 } 15368 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom(byte[] data) 15369 throws com.google.protobuf.InvalidProtocolBufferException { 15370 return PARSER.parseFrom(data); 15371 } 15372 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15373 byte[] data, 15374 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15375 throws com.google.protobuf.InvalidProtocolBufferException { 15376 return PARSER.parseFrom(data, extensionRegistry); 15377 } 15378 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom(java.io.InputStream input) 15379 throws java.io.IOException { 15380 return PARSER.parseFrom(input); 15381 } 15382 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15383 java.io.InputStream input, 15384 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15385 throws java.io.IOException { 15386 return PARSER.parseFrom(input, extensionRegistry); 15387 } 15388 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseDelimitedFrom(java.io.InputStream input) 15389 throws java.io.IOException { 15390 return PARSER.parseDelimitedFrom(input); 15391 } 15392 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseDelimitedFrom( 15393 java.io.InputStream input, 15394 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15395 throws java.io.IOException { 15396 return PARSER.parseDelimitedFrom(input, extensionRegistry); 15397 } 15398 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15399 com.google.protobuf.CodedInputStream input) 15400 throws java.io.IOException { 15401 return PARSER.parseFrom(input); 15402 } 15403 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 15404 com.google.protobuf.CodedInputStream input, 15405 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15406 throws java.io.IOException { 15407 return PARSER.parseFrom(input, extensionRegistry); 15408 } 15409 15410 public static Builder newBuilder() { return Builder.create(); } 15411 public Builder newBuilderForType() { return newBuilder(); } 15412 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff prototype) { 15413 return newBuilder().mergeFrom(prototype); 15414 } 15415 public Builder toBuilder() { return newBuilder(this); } 15416 15417 @java.lang.Override 15418 protected Builder newBuilderForType( 15419 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15420 Builder builder = new Builder(parent); 15421 return builder; 15422 } 15423 /** 15424 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff} 15425 */ 15426 public static final class Builder extends 15427 com.google.protobuf.GeneratedMessage.Builder<Builder> 15428 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiffOrBuilder { 15429 public static final com.google.protobuf.Descriptors.Descriptor 15430 getDescriptor() { 15431 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 15432 } 15433 15434 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15435 internalGetFieldAccessorTable() { 15436 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable 15437 .ensureFieldAccessorsInitialized( 15438 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.Builder.class); 15439 } 15440 15441 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.newBuilder() 15442 private Builder() { 15443 maybeForceBuilderInitialization(); 15444 } 15445 15446 private Builder( 15447 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15448 super(parent); 15449 maybeForceBuilderInitialization(); 15450 } 15451 private void maybeForceBuilderInitialization() { 15452 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 15453 getSnapshotCopyFieldBuilder(); 15454 } 15455 } 15456 private static Builder create() { 15457 return new Builder(); 15458 } 15459 15460 public Builder clear() { 15461 super.clear(); 15462 snapshotId_ = 0; 15463 bitField0_ = (bitField0_ & ~0x00000001); 15464 fileSize_ = 0L; 15465 bitField0_ = (bitField0_ & ~0x00000002); 15466 name_ = com.google.protobuf.ByteString.EMPTY; 15467 bitField0_ = (bitField0_ & ~0x00000004); 15468 if (snapshotCopyBuilder_ == null) { 15469 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 15470 } else { 15471 snapshotCopyBuilder_.clear(); 15472 } 15473 bitField0_ = (bitField0_ & ~0x00000008); 15474 return this; 15475 } 15476 15477 public Builder clone() { 15478 return create().mergeFrom(buildPartial()); 15479 } 15480 15481 public com.google.protobuf.Descriptors.Descriptor 15482 getDescriptorForType() { 15483 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 15484 } 15485 15486 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff getDefaultInstanceForType() { 15487 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.getDefaultInstance(); 15488 } 15489 15490 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff build() { 15491 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff result = buildPartial(); 15492 if (!result.isInitialized()) { 15493 throw newUninitializedMessageException(result); 15494 } 15495 return result; 15496 } 15497 15498 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff buildPartial() { 15499 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff(this); 15500 int from_bitField0_ = bitField0_; 15501 int to_bitField0_ = 0; 15502 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 15503 to_bitField0_ |= 0x00000001; 15504 } 15505 result.snapshotId_ = snapshotId_; 15506 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 15507 to_bitField0_ |= 0x00000002; 15508 } 15509 result.fileSize_ = fileSize_; 15510 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 15511 to_bitField0_ |= 0x00000004; 15512 } 15513 result.name_ = name_; 15514 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 15515 to_bitField0_ |= 0x00000008; 15516 } 15517 if (snapshotCopyBuilder_ == null) { 15518 result.snapshotCopy_ = snapshotCopy_; 15519 } else { 15520 result.snapshotCopy_ = snapshotCopyBuilder_.build(); 15521 } 15522 result.bitField0_ = to_bitField0_; 15523 onBuilt(); 15524 return result; 15525 } 15526 15527 public Builder mergeFrom(com.google.protobuf.Message other) { 15528 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff) { 15529 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff)other); 15530 } else { 15531 super.mergeFrom(other); 15532 return this; 15533 } 15534 } 15535 15536 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff other) { 15537 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.getDefaultInstance()) return this; 15538 if (other.hasSnapshotId()) { 15539 setSnapshotId(other.getSnapshotId()); 15540 } 15541 if (other.hasFileSize()) { 15542 setFileSize(other.getFileSize()); 15543 } 15544 if (other.hasName()) { 15545 setName(other.getName()); 15546 } 15547 if (other.hasSnapshotCopy()) { 15548 mergeSnapshotCopy(other.getSnapshotCopy()); 15549 } 15550 this.mergeUnknownFields(other.getUnknownFields()); 15551 return this; 15552 } 15553 15554 public final boolean isInitialized() { 15555 if (hasSnapshotCopy()) { 15556 if (!getSnapshotCopy().isInitialized()) { 15557 15558 return false; 15559 } 15560 } 15561 return true; 15562 } 15563 15564 public Builder mergeFrom( 15565 com.google.protobuf.CodedInputStream input, 15566 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15567 throws java.io.IOException { 15568 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parsedMessage = null; 15569 try { 15570 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 15571 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15572 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff) e.getUnfinishedMessage(); 15573 throw e; 15574 } finally { 15575 if (parsedMessage != null) { 15576 mergeFrom(parsedMessage); 15577 } 15578 } 15579 return this; 15580 } 15581 private int bitField0_; 15582 15583 // optional uint32 snapshotId = 1; 15584 private int snapshotId_ ; 15585 /** 15586 * <code>optional uint32 snapshotId = 1;</code> 15587 */ 15588 public boolean hasSnapshotId() { 15589 return ((bitField0_ & 0x00000001) == 0x00000001); 15590 } 15591 /** 15592 * <code>optional uint32 snapshotId = 1;</code> 15593 */ 15594 public int getSnapshotId() { 15595 return snapshotId_; 15596 } 15597 /** 15598 * <code>optional uint32 snapshotId = 1;</code> 15599 */ 15600 public Builder setSnapshotId(int value) { 15601 bitField0_ |= 0x00000001; 15602 snapshotId_ = value; 15603 onChanged(); 15604 return this; 15605 } 15606 /** 15607 * <code>optional uint32 snapshotId = 1;</code> 15608 */ 15609 public Builder clearSnapshotId() { 15610 bitField0_ = (bitField0_ & ~0x00000001); 15611 snapshotId_ = 0; 15612 onChanged(); 15613 return this; 15614 } 15615 15616 // optional uint64 fileSize = 2; 15617 private long fileSize_ ; 15618 /** 15619 * <code>optional uint64 fileSize = 2;</code> 15620 */ 15621 public boolean hasFileSize() { 15622 return ((bitField0_ & 0x00000002) == 0x00000002); 15623 } 15624 /** 15625 * <code>optional uint64 fileSize = 2;</code> 15626 */ 15627 public long getFileSize() { 15628 return fileSize_; 15629 } 15630 /** 15631 * <code>optional uint64 fileSize = 2;</code> 15632 */ 15633 public Builder setFileSize(long value) { 15634 bitField0_ |= 0x00000002; 15635 fileSize_ = value; 15636 onChanged(); 15637 return this; 15638 } 15639 /** 15640 * <code>optional uint64 fileSize = 2;</code> 15641 */ 15642 public Builder clearFileSize() { 15643 bitField0_ = (bitField0_ & ~0x00000002); 15644 fileSize_ = 0L; 15645 onChanged(); 15646 return this; 15647 } 15648 15649 // optional bytes name = 3; 15650 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 15651 /** 15652 * <code>optional bytes name = 3;</code> 15653 */ 15654 public boolean hasName() { 15655 return ((bitField0_ & 0x00000004) == 0x00000004); 15656 } 15657 /** 15658 * <code>optional bytes name = 3;</code> 15659 */ 15660 public com.google.protobuf.ByteString getName() { 15661 return name_; 15662 } 15663 /** 15664 * <code>optional bytes name = 3;</code> 15665 */ 15666 public Builder setName(com.google.protobuf.ByteString value) { 15667 if (value == null) { 15668 throw new NullPointerException(); 15669 } 15670 bitField0_ |= 0x00000004; 15671 name_ = value; 15672 onChanged(); 15673 return this; 15674 } 15675 /** 15676 * <code>optional bytes name = 3;</code> 15677 */ 15678 public Builder clearName() { 15679 bitField0_ = (bitField0_ & ~0x00000004); 15680 name_ = getDefaultInstance().getName(); 15681 onChanged(); 15682 return this; 15683 } 15684 15685 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 15686 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 15687 private com.google.protobuf.SingleFieldBuilder< 15688 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> snapshotCopyBuilder_; 15689 /** 15690 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15691 */ 15692 public boolean hasSnapshotCopy() { 15693 return ((bitField0_ & 0x00000008) == 0x00000008); 15694 } 15695 /** 15696 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15697 */ 15698 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy() { 15699 if (snapshotCopyBuilder_ == null) { 15700 return snapshotCopy_; 15701 } else { 15702 return snapshotCopyBuilder_.getMessage(); 15703 } 15704 } 15705 /** 15706 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15707 */ 15708 public Builder setSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 15709 if (snapshotCopyBuilder_ == null) { 15710 if (value == null) { 15711 throw new NullPointerException(); 15712 } 15713 snapshotCopy_ = value; 15714 onChanged(); 15715 } else { 15716 snapshotCopyBuilder_.setMessage(value); 15717 } 15718 bitField0_ |= 0x00000008; 15719 return this; 15720 } 15721 /** 15722 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15723 */ 15724 public Builder setSnapshotCopy( 15725 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder builderForValue) { 15726 if (snapshotCopyBuilder_ == null) { 15727 snapshotCopy_ = builderForValue.build(); 15728 onChanged(); 15729 } else { 15730 snapshotCopyBuilder_.setMessage(builderForValue.build()); 15731 } 15732 bitField0_ |= 0x00000008; 15733 return this; 15734 } 15735 /** 15736 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15737 */ 15738 public Builder mergeSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 15739 if (snapshotCopyBuilder_ == null) { 15740 if (((bitField0_ & 0x00000008) == 0x00000008) && 15741 snapshotCopy_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) { 15742 snapshotCopy_ = 15743 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder(snapshotCopy_).mergeFrom(value).buildPartial(); 15744 } else { 15745 snapshotCopy_ = value; 15746 } 15747 onChanged(); 15748 } else { 15749 snapshotCopyBuilder_.mergeFrom(value); 15750 } 15751 bitField0_ |= 0x00000008; 15752 return this; 15753 } 15754 /** 15755 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15756 */ 15757 public Builder clearSnapshotCopy() { 15758 if (snapshotCopyBuilder_ == null) { 15759 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 15760 onChanged(); 15761 } else { 15762 snapshotCopyBuilder_.clear(); 15763 } 15764 bitField0_ = (bitField0_ & ~0x00000008); 15765 return this; 15766 } 15767 /** 15768 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15769 */ 15770 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder getSnapshotCopyBuilder() { 15771 bitField0_ |= 0x00000008; 15772 onChanged(); 15773 return getSnapshotCopyFieldBuilder().getBuilder(); 15774 } 15775 /** 15776 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15777 */ 15778 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder() { 15779 if (snapshotCopyBuilder_ != null) { 15780 return snapshotCopyBuilder_.getMessageOrBuilder(); 15781 } else { 15782 return snapshotCopy_; 15783 } 15784 } 15785 /** 15786 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 15787 */ 15788 private com.google.protobuf.SingleFieldBuilder< 15789 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> 15790 getSnapshotCopyFieldBuilder() { 15791 if (snapshotCopyBuilder_ == null) { 15792 snapshotCopyBuilder_ = new com.google.protobuf.SingleFieldBuilder< 15793 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder>( 15794 snapshotCopy_, 15795 getParentForChildren(), 15796 isClean()); 15797 snapshotCopy_ = null; 15798 } 15799 return snapshotCopyBuilder_; 15800 } 15801 15802 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff) 15803 } 15804 15805 static { 15806 defaultInstance = new FileDiff(true); 15807 defaultInstance.initFields(); 15808 } 15809 15810 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff) 15811 } 15812 15813 public interface DiffEntryOrBuilder 15814 extends com.google.protobuf.MessageOrBuilder { 15815 15816 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 15817 /** 15818 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 15819 */ 15820 boolean hasType(); 15821 /** 15822 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 15823 */ 15824 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType(); 15825 15826 // optional uint64 inodeId = 2; 15827 /** 15828 * <code>optional uint64 inodeId = 2;</code> 15829 */ 15830 boolean hasInodeId(); 15831 /** 15832 * <code>optional uint64 inodeId = 2;</code> 15833 */ 15834 long getInodeId(); 15835 15836 // optional uint32 numOfDiff = 3; 15837 /** 15838 * <code>optional uint32 numOfDiff = 3;</code> 15839 */ 15840 boolean hasNumOfDiff(); 15841 /** 15842 * <code>optional uint32 numOfDiff = 3;</code> 15843 */ 15844 int getNumOfDiff(); 15845 } 15846 /** 15847 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry} 15848 */ 15849 public static final class DiffEntry extends 15850 com.google.protobuf.GeneratedMessage 15851 implements DiffEntryOrBuilder { 15852 // Use DiffEntry.newBuilder() to construct. 15853 private DiffEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 15854 super(builder); 15855 this.unknownFields = builder.getUnknownFields(); 15856 } 15857 private DiffEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 15858 15859 private static final DiffEntry defaultInstance; 15860 public static DiffEntry getDefaultInstance() { 15861 return defaultInstance; 15862 } 15863 15864 public DiffEntry getDefaultInstanceForType() { 15865 return defaultInstance; 15866 } 15867 15868 private final com.google.protobuf.UnknownFieldSet unknownFields; 15869 @java.lang.Override 15870 public final com.google.protobuf.UnknownFieldSet 15871 getUnknownFields() { 15872 return this.unknownFields; 15873 } 15874 private DiffEntry( 15875 com.google.protobuf.CodedInputStream input, 15876 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15877 throws com.google.protobuf.InvalidProtocolBufferException { 15878 initFields(); 15879 int mutable_bitField0_ = 0; 15880 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15881 com.google.protobuf.UnknownFieldSet.newBuilder(); 15882 try { 15883 boolean done = false; 15884 while (!done) { 15885 int tag = input.readTag(); 15886 switch (tag) { 15887 case 0: 15888 done = true; 15889 break; 15890 default: { 15891 if (!parseUnknownField(input, unknownFields, 15892 extensionRegistry, tag)) { 15893 done = true; 15894 } 15895 break; 15896 } 15897 case 8: { 15898 int rawValue = input.readEnum(); 15899 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type value = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.valueOf(rawValue); 15900 if (value == null) { 15901 unknownFields.mergeVarintField(1, rawValue); 15902 } else { 15903 bitField0_ |= 0x00000001; 15904 type_ = value; 15905 } 15906 break; 15907 } 15908 case 16: { 15909 bitField0_ |= 0x00000002; 15910 inodeId_ = input.readUInt64(); 15911 break; 15912 } 15913 case 24: { 15914 bitField0_ |= 0x00000004; 15915 numOfDiff_ = input.readUInt32(); 15916 break; 15917 } 15918 } 15919 } 15920 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15921 throw e.setUnfinishedMessage(this); 15922 } catch (java.io.IOException e) { 15923 throw new com.google.protobuf.InvalidProtocolBufferException( 15924 e.getMessage()).setUnfinishedMessage(this); 15925 } finally { 15926 this.unknownFields = unknownFields.build(); 15927 makeExtensionsImmutable(); 15928 } 15929 } 15930 public static final com.google.protobuf.Descriptors.Descriptor 15931 getDescriptor() { 15932 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 15933 } 15934 15935 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15936 internalGetFieldAccessorTable() { 15937 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable 15938 .ensureFieldAccessorsInitialized( 15939 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Builder.class); 15940 } 15941 15942 public static com.google.protobuf.Parser<DiffEntry> PARSER = 15943 new com.google.protobuf.AbstractParser<DiffEntry>() { 15944 public DiffEntry parsePartialFrom( 15945 com.google.protobuf.CodedInputStream input, 15946 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15947 throws com.google.protobuf.InvalidProtocolBufferException { 15948 return new DiffEntry(input, extensionRegistry); 15949 } 15950 }; 15951 15952 @java.lang.Override 15953 public com.google.protobuf.Parser<DiffEntry> getParserForType() { 15954 return PARSER; 15955 } 15956 15957 /** 15958 * Protobuf enum {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type} 15959 */ 15960 public enum Type 15961 implements com.google.protobuf.ProtocolMessageEnum { 15962 /** 15963 * <code>FILEDIFF = 1;</code> 15964 */ 15965 FILEDIFF(0, 1), 15966 /** 15967 * <code>DIRECTORYDIFF = 2;</code> 15968 */ 15969 DIRECTORYDIFF(1, 2), 15970 ; 15971 15972 /** 15973 * <code>FILEDIFF = 1;</code> 15974 */ 15975 public static final int FILEDIFF_VALUE = 1; 15976 /** 15977 * <code>DIRECTORYDIFF = 2;</code> 15978 */ 15979 public static final int DIRECTORYDIFF_VALUE = 2; 15980 15981 15982 public final int getNumber() { return value; } 15983 15984 public static Type valueOf(int value) { 15985 switch (value) { 15986 case 1: return FILEDIFF; 15987 case 2: return DIRECTORYDIFF; 15988 default: return null; 15989 } 15990 } 15991 15992 public static com.google.protobuf.Internal.EnumLiteMap<Type> 15993 internalGetValueMap() { 15994 return internalValueMap; 15995 } 15996 private static com.google.protobuf.Internal.EnumLiteMap<Type> 15997 internalValueMap = 15998 new com.google.protobuf.Internal.EnumLiteMap<Type>() { 15999 public Type findValueByNumber(int number) { 16000 return Type.valueOf(number); 16001 } 16002 }; 16003 16004 public final com.google.protobuf.Descriptors.EnumValueDescriptor 16005 getValueDescriptor() { 16006 return getDescriptor().getValues().get(index); 16007 } 16008 public final com.google.protobuf.Descriptors.EnumDescriptor 16009 getDescriptorForType() { 16010 return getDescriptor(); 16011 } 16012 public static final com.google.protobuf.Descriptors.EnumDescriptor 16013 getDescriptor() { 16014 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDescriptor().getEnumTypes().get(0); 16015 } 16016 16017 private static final Type[] VALUES = values(); 16018 16019 public static Type valueOf( 16020 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 16021 if (desc.getType() != getDescriptor()) { 16022 throw new java.lang.IllegalArgumentException( 16023 "EnumValueDescriptor is not for this type."); 16024 } 16025 return VALUES[desc.getIndex()]; 16026 } 16027 16028 private final int index; 16029 private final int value; 16030 16031 private Type(int index, int value) { 16032 this.index = index; 16033 this.value = value; 16034 } 16035 16036 // @@protoc_insertion_point(enum_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type) 16037 } 16038 16039 private int bitField0_; 16040 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 16041 public static final int TYPE_FIELD_NUMBER = 1; 16042 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type type_; 16043 /** 16044 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16045 */ 16046 public boolean hasType() { 16047 return ((bitField0_ & 0x00000001) == 0x00000001); 16048 } 16049 /** 16050 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16051 */ 16052 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType() { 16053 return type_; 16054 } 16055 16056 // optional uint64 inodeId = 2; 16057 public static final int INODEID_FIELD_NUMBER = 2; 16058 private long inodeId_; 16059 /** 16060 * <code>optional uint64 inodeId = 2;</code> 16061 */ 16062 public boolean hasInodeId() { 16063 return ((bitField0_ & 0x00000002) == 0x00000002); 16064 } 16065 /** 16066 * <code>optional uint64 inodeId = 2;</code> 16067 */ 16068 public long getInodeId() { 16069 return inodeId_; 16070 } 16071 16072 // optional uint32 numOfDiff = 3; 16073 public static final int NUMOFDIFF_FIELD_NUMBER = 3; 16074 private int numOfDiff_; 16075 /** 16076 * <code>optional uint32 numOfDiff = 3;</code> 16077 */ 16078 public boolean hasNumOfDiff() { 16079 return ((bitField0_ & 0x00000004) == 0x00000004); 16080 } 16081 /** 16082 * <code>optional uint32 numOfDiff = 3;</code> 16083 */ 16084 public int getNumOfDiff() { 16085 return numOfDiff_; 16086 } 16087 16088 private void initFields() { 16089 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 16090 inodeId_ = 0L; 16091 numOfDiff_ = 0; 16092 } 16093 private byte memoizedIsInitialized = -1; 16094 public final boolean isInitialized() { 16095 byte isInitialized = memoizedIsInitialized; 16096 if (isInitialized != -1) return isInitialized == 1; 16097 16098 if (!hasType()) { 16099 memoizedIsInitialized = 0; 16100 return false; 16101 } 16102 memoizedIsInitialized = 1; 16103 return true; 16104 } 16105 16106 public void writeTo(com.google.protobuf.CodedOutputStream output) 16107 throws java.io.IOException { 16108 getSerializedSize(); 16109 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16110 output.writeEnum(1, type_.getNumber()); 16111 } 16112 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16113 output.writeUInt64(2, inodeId_); 16114 } 16115 if (((bitField0_ & 0x00000004) == 0x00000004)) { 16116 output.writeUInt32(3, numOfDiff_); 16117 } 16118 getUnknownFields().writeTo(output); 16119 } 16120 16121 private int memoizedSerializedSize = -1; 16122 public int getSerializedSize() { 16123 int size = memoizedSerializedSize; 16124 if (size != -1) return size; 16125 16126 size = 0; 16127 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16128 size += com.google.protobuf.CodedOutputStream 16129 .computeEnumSize(1, type_.getNumber()); 16130 } 16131 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16132 size += com.google.protobuf.CodedOutputStream 16133 .computeUInt64Size(2, inodeId_); 16134 } 16135 if (((bitField0_ & 0x00000004) == 0x00000004)) { 16136 size += com.google.protobuf.CodedOutputStream 16137 .computeUInt32Size(3, numOfDiff_); 16138 } 16139 size += getUnknownFields().getSerializedSize(); 16140 memoizedSerializedSize = size; 16141 return size; 16142 } 16143 16144 private static final long serialVersionUID = 0L; 16145 @java.lang.Override 16146 protected java.lang.Object writeReplace() 16147 throws java.io.ObjectStreamException { 16148 return super.writeReplace(); 16149 } 16150 16151 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16152 com.google.protobuf.ByteString data) 16153 throws com.google.protobuf.InvalidProtocolBufferException { 16154 return PARSER.parseFrom(data); 16155 } 16156 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16157 com.google.protobuf.ByteString data, 16158 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16159 throws com.google.protobuf.InvalidProtocolBufferException { 16160 return PARSER.parseFrom(data, extensionRegistry); 16161 } 16162 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom(byte[] data) 16163 throws com.google.protobuf.InvalidProtocolBufferException { 16164 return PARSER.parseFrom(data); 16165 } 16166 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16167 byte[] data, 16168 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16169 throws com.google.protobuf.InvalidProtocolBufferException { 16170 return PARSER.parseFrom(data, extensionRegistry); 16171 } 16172 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom(java.io.InputStream input) 16173 throws java.io.IOException { 16174 return PARSER.parseFrom(input); 16175 } 16176 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16177 java.io.InputStream input, 16178 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16179 throws java.io.IOException { 16180 return PARSER.parseFrom(input, extensionRegistry); 16181 } 16182 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseDelimitedFrom(java.io.InputStream input) 16183 throws java.io.IOException { 16184 return PARSER.parseDelimitedFrom(input); 16185 } 16186 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseDelimitedFrom( 16187 java.io.InputStream input, 16188 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16189 throws java.io.IOException { 16190 return PARSER.parseDelimitedFrom(input, extensionRegistry); 16191 } 16192 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16193 com.google.protobuf.CodedInputStream input) 16194 throws java.io.IOException { 16195 return PARSER.parseFrom(input); 16196 } 16197 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 16198 com.google.protobuf.CodedInputStream input, 16199 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16200 throws java.io.IOException { 16201 return PARSER.parseFrom(input, extensionRegistry); 16202 } 16203 16204 public static Builder newBuilder() { return Builder.create(); } 16205 public Builder newBuilderForType() { return newBuilder(); } 16206 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry prototype) { 16207 return newBuilder().mergeFrom(prototype); 16208 } 16209 public Builder toBuilder() { return newBuilder(this); } 16210 16211 @java.lang.Override 16212 protected Builder newBuilderForType( 16213 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16214 Builder builder = new Builder(parent); 16215 return builder; 16216 } 16217 /** 16218 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry} 16219 */ 16220 public static final class Builder extends 16221 com.google.protobuf.GeneratedMessage.Builder<Builder> 16222 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntryOrBuilder { 16223 public static final com.google.protobuf.Descriptors.Descriptor 16224 getDescriptor() { 16225 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 16226 } 16227 16228 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16229 internalGetFieldAccessorTable() { 16230 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable 16231 .ensureFieldAccessorsInitialized( 16232 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Builder.class); 16233 } 16234 16235 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.newBuilder() 16236 private Builder() { 16237 maybeForceBuilderInitialization(); 16238 } 16239 16240 private Builder( 16241 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16242 super(parent); 16243 maybeForceBuilderInitialization(); 16244 } 16245 private void maybeForceBuilderInitialization() { 16246 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 16247 } 16248 } 16249 private static Builder create() { 16250 return new Builder(); 16251 } 16252 16253 public Builder clear() { 16254 super.clear(); 16255 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 16256 bitField0_ = (bitField0_ & ~0x00000001); 16257 inodeId_ = 0L; 16258 bitField0_ = (bitField0_ & ~0x00000002); 16259 numOfDiff_ = 0; 16260 bitField0_ = (bitField0_ & ~0x00000004); 16261 return this; 16262 } 16263 16264 public Builder clone() { 16265 return create().mergeFrom(buildPartial()); 16266 } 16267 16268 public com.google.protobuf.Descriptors.Descriptor 16269 getDescriptorForType() { 16270 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 16271 } 16272 16273 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry getDefaultInstanceForType() { 16274 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDefaultInstance(); 16275 } 16276 16277 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry build() { 16278 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry result = buildPartial(); 16279 if (!result.isInitialized()) { 16280 throw newUninitializedMessageException(result); 16281 } 16282 return result; 16283 } 16284 16285 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry buildPartial() { 16286 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry(this); 16287 int from_bitField0_ = bitField0_; 16288 int to_bitField0_ = 0; 16289 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 16290 to_bitField0_ |= 0x00000001; 16291 } 16292 result.type_ = type_; 16293 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 16294 to_bitField0_ |= 0x00000002; 16295 } 16296 result.inodeId_ = inodeId_; 16297 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 16298 to_bitField0_ |= 0x00000004; 16299 } 16300 result.numOfDiff_ = numOfDiff_; 16301 result.bitField0_ = to_bitField0_; 16302 onBuilt(); 16303 return result; 16304 } 16305 16306 public Builder mergeFrom(com.google.protobuf.Message other) { 16307 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry) { 16308 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry)other); 16309 } else { 16310 super.mergeFrom(other); 16311 return this; 16312 } 16313 } 16314 16315 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry other) { 16316 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDefaultInstance()) return this; 16317 if (other.hasType()) { 16318 setType(other.getType()); 16319 } 16320 if (other.hasInodeId()) { 16321 setInodeId(other.getInodeId()); 16322 } 16323 if (other.hasNumOfDiff()) { 16324 setNumOfDiff(other.getNumOfDiff()); 16325 } 16326 this.mergeUnknownFields(other.getUnknownFields()); 16327 return this; 16328 } 16329 16330 public final boolean isInitialized() { 16331 if (!hasType()) { 16332 16333 return false; 16334 } 16335 return true; 16336 } 16337 16338 public Builder mergeFrom( 16339 com.google.protobuf.CodedInputStream input, 16340 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16341 throws java.io.IOException { 16342 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parsedMessage = null; 16343 try { 16344 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 16345 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16346 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry) e.getUnfinishedMessage(); 16347 throw e; 16348 } finally { 16349 if (parsedMessage != null) { 16350 mergeFrom(parsedMessage); 16351 } 16352 } 16353 return this; 16354 } 16355 private int bitField0_; 16356 16357 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 16358 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 16359 /** 16360 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16361 */ 16362 public boolean hasType() { 16363 return ((bitField0_ & 0x00000001) == 0x00000001); 16364 } 16365 /** 16366 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16367 */ 16368 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType() { 16369 return type_; 16370 } 16371 /** 16372 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16373 */ 16374 public Builder setType(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type value) { 16375 if (value == null) { 16376 throw new NullPointerException(); 16377 } 16378 bitField0_ |= 0x00000001; 16379 type_ = value; 16380 onChanged(); 16381 return this; 16382 } 16383 /** 16384 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 16385 */ 16386 public Builder clearType() { 16387 bitField0_ = (bitField0_ & ~0x00000001); 16388 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 16389 onChanged(); 16390 return this; 16391 } 16392 16393 // optional uint64 inodeId = 2; 16394 private long inodeId_ ; 16395 /** 16396 * <code>optional uint64 inodeId = 2;</code> 16397 */ 16398 public boolean hasInodeId() { 16399 return ((bitField0_ & 0x00000002) == 0x00000002); 16400 } 16401 /** 16402 * <code>optional uint64 inodeId = 2;</code> 16403 */ 16404 public long getInodeId() { 16405 return inodeId_; 16406 } 16407 /** 16408 * <code>optional uint64 inodeId = 2;</code> 16409 */ 16410 public Builder setInodeId(long value) { 16411 bitField0_ |= 0x00000002; 16412 inodeId_ = value; 16413 onChanged(); 16414 return this; 16415 } 16416 /** 16417 * <code>optional uint64 inodeId = 2;</code> 16418 */ 16419 public Builder clearInodeId() { 16420 bitField0_ = (bitField0_ & ~0x00000002); 16421 inodeId_ = 0L; 16422 onChanged(); 16423 return this; 16424 } 16425 16426 // optional uint32 numOfDiff = 3; 16427 private int numOfDiff_ ; 16428 /** 16429 * <code>optional uint32 numOfDiff = 3;</code> 16430 */ 16431 public boolean hasNumOfDiff() { 16432 return ((bitField0_ & 0x00000004) == 0x00000004); 16433 } 16434 /** 16435 * <code>optional uint32 numOfDiff = 3;</code> 16436 */ 16437 public int getNumOfDiff() { 16438 return numOfDiff_; 16439 } 16440 /** 16441 * <code>optional uint32 numOfDiff = 3;</code> 16442 */ 16443 public Builder setNumOfDiff(int value) { 16444 bitField0_ |= 0x00000004; 16445 numOfDiff_ = value; 16446 onChanged(); 16447 return this; 16448 } 16449 /** 16450 * <code>optional uint32 numOfDiff = 3;</code> 16451 */ 16452 public Builder clearNumOfDiff() { 16453 bitField0_ = (bitField0_ & ~0x00000004); 16454 numOfDiff_ = 0; 16455 onChanged(); 16456 return this; 16457 } 16458 16459 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry) 16460 } 16461 16462 static { 16463 defaultInstance = new DiffEntry(true); 16464 defaultInstance.initFields(); 16465 } 16466 16467 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry) 16468 } 16469 16470 private void initFields() { 16471 } 16472 private byte memoizedIsInitialized = -1; 16473 public final boolean isInitialized() { 16474 byte isInitialized = memoizedIsInitialized; 16475 if (isInitialized != -1) return isInitialized == 1; 16476 16477 memoizedIsInitialized = 1; 16478 return true; 16479 } 16480 16481 public void writeTo(com.google.protobuf.CodedOutputStream output) 16482 throws java.io.IOException { 16483 getSerializedSize(); 16484 getUnknownFields().writeTo(output); 16485 } 16486 16487 private int memoizedSerializedSize = -1; 16488 public int getSerializedSize() { 16489 int size = memoizedSerializedSize; 16490 if (size != -1) return size; 16491 16492 size = 0; 16493 size += getUnknownFields().getSerializedSize(); 16494 memoizedSerializedSize = size; 16495 return size; 16496 } 16497 16498 private static final long serialVersionUID = 0L; 16499 @java.lang.Override 16500 protected java.lang.Object writeReplace() 16501 throws java.io.ObjectStreamException { 16502 return super.writeReplace(); 16503 } 16504 16505 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16506 com.google.protobuf.ByteString data) 16507 throws com.google.protobuf.InvalidProtocolBufferException { 16508 return PARSER.parseFrom(data); 16509 } 16510 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16511 com.google.protobuf.ByteString data, 16512 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16513 throws com.google.protobuf.InvalidProtocolBufferException { 16514 return PARSER.parseFrom(data, extensionRegistry); 16515 } 16516 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom(byte[] data) 16517 throws com.google.protobuf.InvalidProtocolBufferException { 16518 return PARSER.parseFrom(data); 16519 } 16520 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16521 byte[] data, 16522 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16523 throws com.google.protobuf.InvalidProtocolBufferException { 16524 return PARSER.parseFrom(data, extensionRegistry); 16525 } 16526 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom(java.io.InputStream input) 16527 throws java.io.IOException { 16528 return PARSER.parseFrom(input); 16529 } 16530 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16531 java.io.InputStream input, 16532 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16533 throws java.io.IOException { 16534 return PARSER.parseFrom(input, extensionRegistry); 16535 } 16536 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseDelimitedFrom(java.io.InputStream input) 16537 throws java.io.IOException { 16538 return PARSER.parseDelimitedFrom(input); 16539 } 16540 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseDelimitedFrom( 16541 java.io.InputStream input, 16542 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16543 throws java.io.IOException { 16544 return PARSER.parseDelimitedFrom(input, extensionRegistry); 16545 } 16546 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16547 com.google.protobuf.CodedInputStream input) 16548 throws java.io.IOException { 16549 return PARSER.parseFrom(input); 16550 } 16551 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 16552 com.google.protobuf.CodedInputStream input, 16553 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16554 throws java.io.IOException { 16555 return PARSER.parseFrom(input, extensionRegistry); 16556 } 16557 16558 public static Builder newBuilder() { return Builder.create(); } 16559 public Builder newBuilderForType() { return newBuilder(); } 16560 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection prototype) { 16561 return newBuilder().mergeFrom(prototype); 16562 } 16563 public Builder toBuilder() { return newBuilder(this); } 16564 16565 @java.lang.Override 16566 protected Builder newBuilderForType( 16567 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16568 Builder builder = new Builder(parent); 16569 return builder; 16570 } 16571 /** 16572 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection} 16573 * 16574 * <pre> 16575 ** 16576 * This section records information about snapshot diffs 16577 * NAME: SNAPSHOT_DIFF 16578 * </pre> 16579 */ 16580 public static final class Builder extends 16581 com.google.protobuf.GeneratedMessage.Builder<Builder> 16582 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSectionOrBuilder { 16583 public static final com.google.protobuf.Descriptors.Descriptor 16584 getDescriptor() { 16585 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 16586 } 16587 16588 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16589 internalGetFieldAccessorTable() { 16590 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable 16591 .ensureFieldAccessorsInitialized( 16592 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.Builder.class); 16593 } 16594 16595 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.newBuilder() 16596 private Builder() { 16597 maybeForceBuilderInitialization(); 16598 } 16599 16600 private Builder( 16601 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16602 super(parent); 16603 maybeForceBuilderInitialization(); 16604 } 16605 private void maybeForceBuilderInitialization() { 16606 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 16607 } 16608 } 16609 private static Builder create() { 16610 return new Builder(); 16611 } 16612 16613 public Builder clear() { 16614 super.clear(); 16615 return this; 16616 } 16617 16618 public Builder clone() { 16619 return create().mergeFrom(buildPartial()); 16620 } 16621 16622 public com.google.protobuf.Descriptors.Descriptor 16623 getDescriptorForType() { 16624 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 16625 } 16626 16627 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection getDefaultInstanceForType() { 16628 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.getDefaultInstance(); 16629 } 16630 16631 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection build() { 16632 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection result = buildPartial(); 16633 if (!result.isInitialized()) { 16634 throw newUninitializedMessageException(result); 16635 } 16636 return result; 16637 } 16638 16639 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection buildPartial() { 16640 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection(this); 16641 onBuilt(); 16642 return result; 16643 } 16644 16645 public Builder mergeFrom(com.google.protobuf.Message other) { 16646 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection) { 16647 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection)other); 16648 } else { 16649 super.mergeFrom(other); 16650 return this; 16651 } 16652 } 16653 16654 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection other) { 16655 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.getDefaultInstance()) return this; 16656 this.mergeUnknownFields(other.getUnknownFields()); 16657 return this; 16658 } 16659 16660 public final boolean isInitialized() { 16661 return true; 16662 } 16663 16664 public Builder mergeFrom( 16665 com.google.protobuf.CodedInputStream input, 16666 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16667 throws java.io.IOException { 16668 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parsedMessage = null; 16669 try { 16670 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 16671 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16672 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection) e.getUnfinishedMessage(); 16673 throw e; 16674 } finally { 16675 if (parsedMessage != null) { 16676 mergeFrom(parsedMessage); 16677 } 16678 } 16679 return this; 16680 } 16681 16682 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection) 16683 } 16684 16685 static { 16686 defaultInstance = new SnapshotDiffSection(true); 16687 defaultInstance.initFields(); 16688 } 16689 16690 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection) 16691 } 16692 16693 public interface StringTableSectionOrBuilder 16694 extends com.google.protobuf.MessageOrBuilder { 16695 16696 // optional uint32 numEntry = 1; 16697 /** 16698 * <code>optional uint32 numEntry = 1;</code> 16699 * 16700 * <pre> 16701 * repeated Entry 16702 * </pre> 16703 */ 16704 boolean hasNumEntry(); 16705 /** 16706 * <code>optional uint32 numEntry = 1;</code> 16707 * 16708 * <pre> 16709 * repeated Entry 16710 * </pre> 16711 */ 16712 int getNumEntry(); 16713 } 16714 /** 16715 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection} 16716 * 16717 * <pre> 16718 ** 16719 * This section maps string to id 16720 * NAME: STRING_TABLE 16721 * </pre> 16722 */ 16723 public static final class StringTableSection extends 16724 com.google.protobuf.GeneratedMessage 16725 implements StringTableSectionOrBuilder { 16726 // Use StringTableSection.newBuilder() to construct. 16727 private StringTableSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 16728 super(builder); 16729 this.unknownFields = builder.getUnknownFields(); 16730 } 16731 private StringTableSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 16732 16733 private static final StringTableSection defaultInstance; 16734 public static StringTableSection getDefaultInstance() { 16735 return defaultInstance; 16736 } 16737 16738 public StringTableSection getDefaultInstanceForType() { 16739 return defaultInstance; 16740 } 16741 16742 private final com.google.protobuf.UnknownFieldSet unknownFields; 16743 @java.lang.Override 16744 public final com.google.protobuf.UnknownFieldSet 16745 getUnknownFields() { 16746 return this.unknownFields; 16747 } 16748 private StringTableSection( 16749 com.google.protobuf.CodedInputStream input, 16750 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16751 throws com.google.protobuf.InvalidProtocolBufferException { 16752 initFields(); 16753 int mutable_bitField0_ = 0; 16754 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 16755 com.google.protobuf.UnknownFieldSet.newBuilder(); 16756 try { 16757 boolean done = false; 16758 while (!done) { 16759 int tag = input.readTag(); 16760 switch (tag) { 16761 case 0: 16762 done = true; 16763 break; 16764 default: { 16765 if (!parseUnknownField(input, unknownFields, 16766 extensionRegistry, tag)) { 16767 done = true; 16768 } 16769 break; 16770 } 16771 case 8: { 16772 bitField0_ |= 0x00000001; 16773 numEntry_ = input.readUInt32(); 16774 break; 16775 } 16776 } 16777 } 16778 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16779 throw e.setUnfinishedMessage(this); 16780 } catch (java.io.IOException e) { 16781 throw new com.google.protobuf.InvalidProtocolBufferException( 16782 e.getMessage()).setUnfinishedMessage(this); 16783 } finally { 16784 this.unknownFields = unknownFields.build(); 16785 makeExtensionsImmutable(); 16786 } 16787 } 16788 public static final com.google.protobuf.Descriptors.Descriptor 16789 getDescriptor() { 16790 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 16791 } 16792 16793 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16794 internalGetFieldAccessorTable() { 16795 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable 16796 .ensureFieldAccessorsInitialized( 16797 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Builder.class); 16798 } 16799 16800 public static com.google.protobuf.Parser<StringTableSection> PARSER = 16801 new com.google.protobuf.AbstractParser<StringTableSection>() { 16802 public StringTableSection parsePartialFrom( 16803 com.google.protobuf.CodedInputStream input, 16804 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16805 throws com.google.protobuf.InvalidProtocolBufferException { 16806 return new StringTableSection(input, extensionRegistry); 16807 } 16808 }; 16809 16810 @java.lang.Override 16811 public com.google.protobuf.Parser<StringTableSection> getParserForType() { 16812 return PARSER; 16813 } 16814 16815 public interface EntryOrBuilder 16816 extends com.google.protobuf.MessageOrBuilder { 16817 16818 // optional uint32 id = 1; 16819 /** 16820 * <code>optional uint32 id = 1;</code> 16821 */ 16822 boolean hasId(); 16823 /** 16824 * <code>optional uint32 id = 1;</code> 16825 */ 16826 int getId(); 16827 16828 // optional string str = 2; 16829 /** 16830 * <code>optional string str = 2;</code> 16831 */ 16832 boolean hasStr(); 16833 /** 16834 * <code>optional string str = 2;</code> 16835 */ 16836 java.lang.String getStr(); 16837 /** 16838 * <code>optional string str = 2;</code> 16839 */ 16840 com.google.protobuf.ByteString 16841 getStrBytes(); 16842 } 16843 /** 16844 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection.Entry} 16845 */ 16846 public static final class Entry extends 16847 com.google.protobuf.GeneratedMessage 16848 implements EntryOrBuilder { 16849 // Use Entry.newBuilder() to construct. 16850 private Entry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 16851 super(builder); 16852 this.unknownFields = builder.getUnknownFields(); 16853 } 16854 private Entry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 16855 16856 private static final Entry defaultInstance; 16857 public static Entry getDefaultInstance() { 16858 return defaultInstance; 16859 } 16860 16861 public Entry getDefaultInstanceForType() { 16862 return defaultInstance; 16863 } 16864 16865 private final com.google.protobuf.UnknownFieldSet unknownFields; 16866 @java.lang.Override 16867 public final com.google.protobuf.UnknownFieldSet 16868 getUnknownFields() { 16869 return this.unknownFields; 16870 } 16871 private Entry( 16872 com.google.protobuf.CodedInputStream input, 16873 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16874 throws com.google.protobuf.InvalidProtocolBufferException { 16875 initFields(); 16876 int mutable_bitField0_ = 0; 16877 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 16878 com.google.protobuf.UnknownFieldSet.newBuilder(); 16879 try { 16880 boolean done = false; 16881 while (!done) { 16882 int tag = input.readTag(); 16883 switch (tag) { 16884 case 0: 16885 done = true; 16886 break; 16887 default: { 16888 if (!parseUnknownField(input, unknownFields, 16889 extensionRegistry, tag)) { 16890 done = true; 16891 } 16892 break; 16893 } 16894 case 8: { 16895 bitField0_ |= 0x00000001; 16896 id_ = input.readUInt32(); 16897 break; 16898 } 16899 case 18: { 16900 bitField0_ |= 0x00000002; 16901 str_ = input.readBytes(); 16902 break; 16903 } 16904 } 16905 } 16906 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16907 throw e.setUnfinishedMessage(this); 16908 } catch (java.io.IOException e) { 16909 throw new com.google.protobuf.InvalidProtocolBufferException( 16910 e.getMessage()).setUnfinishedMessage(this); 16911 } finally { 16912 this.unknownFields = unknownFields.build(); 16913 makeExtensionsImmutable(); 16914 } 16915 } 16916 public static final com.google.protobuf.Descriptors.Descriptor 16917 getDescriptor() { 16918 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 16919 } 16920 16921 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16922 internalGetFieldAccessorTable() { 16923 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable 16924 .ensureFieldAccessorsInitialized( 16925 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.Builder.class); 16926 } 16927 16928 public static com.google.protobuf.Parser<Entry> PARSER = 16929 new com.google.protobuf.AbstractParser<Entry>() { 16930 public Entry parsePartialFrom( 16931 com.google.protobuf.CodedInputStream input, 16932 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16933 throws com.google.protobuf.InvalidProtocolBufferException { 16934 return new Entry(input, extensionRegistry); 16935 } 16936 }; 16937 16938 @java.lang.Override 16939 public com.google.protobuf.Parser<Entry> getParserForType() { 16940 return PARSER; 16941 } 16942 16943 private int bitField0_; 16944 // optional uint32 id = 1; 16945 public static final int ID_FIELD_NUMBER = 1; 16946 private int id_; 16947 /** 16948 * <code>optional uint32 id = 1;</code> 16949 */ 16950 public boolean hasId() { 16951 return ((bitField0_ & 0x00000001) == 0x00000001); 16952 } 16953 /** 16954 * <code>optional uint32 id = 1;</code> 16955 */ 16956 public int getId() { 16957 return id_; 16958 } 16959 16960 // optional string str = 2; 16961 public static final int STR_FIELD_NUMBER = 2; 16962 private java.lang.Object str_; 16963 /** 16964 * <code>optional string str = 2;</code> 16965 */ 16966 public boolean hasStr() { 16967 return ((bitField0_ & 0x00000002) == 0x00000002); 16968 } 16969 /** 16970 * <code>optional string str = 2;</code> 16971 */ 16972 public java.lang.String getStr() { 16973 java.lang.Object ref = str_; 16974 if (ref instanceof java.lang.String) { 16975 return (java.lang.String) ref; 16976 } else { 16977 com.google.protobuf.ByteString bs = 16978 (com.google.protobuf.ByteString) ref; 16979 java.lang.String s = bs.toStringUtf8(); 16980 if (bs.isValidUtf8()) { 16981 str_ = s; 16982 } 16983 return s; 16984 } 16985 } 16986 /** 16987 * <code>optional string str = 2;</code> 16988 */ 16989 public com.google.protobuf.ByteString 16990 getStrBytes() { 16991 java.lang.Object ref = str_; 16992 if (ref instanceof java.lang.String) { 16993 com.google.protobuf.ByteString b = 16994 com.google.protobuf.ByteString.copyFromUtf8( 16995 (java.lang.String) ref); 16996 str_ = b; 16997 return b; 16998 } else { 16999 return (com.google.protobuf.ByteString) ref; 17000 } 17001 } 17002 17003 private void initFields() { 17004 id_ = 0; 17005 str_ = ""; 17006 } 17007 private byte memoizedIsInitialized = -1; 17008 public final boolean isInitialized() { 17009 byte isInitialized = memoizedIsInitialized; 17010 if (isInitialized != -1) return isInitialized == 1; 17011 17012 memoizedIsInitialized = 1; 17013 return true; 17014 } 17015 17016 public void writeTo(com.google.protobuf.CodedOutputStream output) 17017 throws java.io.IOException { 17018 getSerializedSize(); 17019 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17020 output.writeUInt32(1, id_); 17021 } 17022 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17023 output.writeBytes(2, getStrBytes()); 17024 } 17025 getUnknownFields().writeTo(output); 17026 } 17027 17028 private int memoizedSerializedSize = -1; 17029 public int getSerializedSize() { 17030 int size = memoizedSerializedSize; 17031 if (size != -1) return size; 17032 17033 size = 0; 17034 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17035 size += com.google.protobuf.CodedOutputStream 17036 .computeUInt32Size(1, id_); 17037 } 17038 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17039 size += com.google.protobuf.CodedOutputStream 17040 .computeBytesSize(2, getStrBytes()); 17041 } 17042 size += getUnknownFields().getSerializedSize(); 17043 memoizedSerializedSize = size; 17044 return size; 17045 } 17046 17047 private static final long serialVersionUID = 0L; 17048 @java.lang.Override 17049 protected java.lang.Object writeReplace() 17050 throws java.io.ObjectStreamException { 17051 return super.writeReplace(); 17052 } 17053 17054 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17055 com.google.protobuf.ByteString data) 17056 throws com.google.protobuf.InvalidProtocolBufferException { 17057 return PARSER.parseFrom(data); 17058 } 17059 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17060 com.google.protobuf.ByteString data, 17061 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17062 throws com.google.protobuf.InvalidProtocolBufferException { 17063 return PARSER.parseFrom(data, extensionRegistry); 17064 } 17065 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom(byte[] data) 17066 throws com.google.protobuf.InvalidProtocolBufferException { 17067 return PARSER.parseFrom(data); 17068 } 17069 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17070 byte[] data, 17071 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17072 throws com.google.protobuf.InvalidProtocolBufferException { 17073 return PARSER.parseFrom(data, extensionRegistry); 17074 } 17075 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom(java.io.InputStream input) 17076 throws java.io.IOException { 17077 return PARSER.parseFrom(input); 17078 } 17079 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17080 java.io.InputStream input, 17081 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17082 throws java.io.IOException { 17083 return PARSER.parseFrom(input, extensionRegistry); 17084 } 17085 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseDelimitedFrom(java.io.InputStream input) 17086 throws java.io.IOException { 17087 return PARSER.parseDelimitedFrom(input); 17088 } 17089 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseDelimitedFrom( 17090 java.io.InputStream input, 17091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17092 throws java.io.IOException { 17093 return PARSER.parseDelimitedFrom(input, extensionRegistry); 17094 } 17095 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17096 com.google.protobuf.CodedInputStream input) 17097 throws java.io.IOException { 17098 return PARSER.parseFrom(input); 17099 } 17100 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 17101 com.google.protobuf.CodedInputStream input, 17102 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17103 throws java.io.IOException { 17104 return PARSER.parseFrom(input, extensionRegistry); 17105 } 17106 17107 public static Builder newBuilder() { return Builder.create(); } 17108 public Builder newBuilderForType() { return newBuilder(); } 17109 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry prototype) { 17110 return newBuilder().mergeFrom(prototype); 17111 } 17112 public Builder toBuilder() { return newBuilder(this); } 17113 17114 @java.lang.Override 17115 protected Builder newBuilderForType( 17116 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17117 Builder builder = new Builder(parent); 17118 return builder; 17119 } 17120 /** 17121 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection.Entry} 17122 */ 17123 public static final class Builder extends 17124 com.google.protobuf.GeneratedMessage.Builder<Builder> 17125 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.EntryOrBuilder { 17126 public static final com.google.protobuf.Descriptors.Descriptor 17127 getDescriptor() { 17128 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 17129 } 17130 17131 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17132 internalGetFieldAccessorTable() { 17133 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable 17134 .ensureFieldAccessorsInitialized( 17135 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.Builder.class); 17136 } 17137 17138 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.newBuilder() 17139 private Builder() { 17140 maybeForceBuilderInitialization(); 17141 } 17142 17143 private Builder( 17144 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17145 super(parent); 17146 maybeForceBuilderInitialization(); 17147 } 17148 private void maybeForceBuilderInitialization() { 17149 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 17150 } 17151 } 17152 private static Builder create() { 17153 return new Builder(); 17154 } 17155 17156 public Builder clear() { 17157 super.clear(); 17158 id_ = 0; 17159 bitField0_ = (bitField0_ & ~0x00000001); 17160 str_ = ""; 17161 bitField0_ = (bitField0_ & ~0x00000002); 17162 return this; 17163 } 17164 17165 public Builder clone() { 17166 return create().mergeFrom(buildPartial()); 17167 } 17168 17169 public com.google.protobuf.Descriptors.Descriptor 17170 getDescriptorForType() { 17171 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 17172 } 17173 17174 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry getDefaultInstanceForType() { 17175 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.getDefaultInstance(); 17176 } 17177 17178 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry build() { 17179 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry result = buildPartial(); 17180 if (!result.isInitialized()) { 17181 throw newUninitializedMessageException(result); 17182 } 17183 return result; 17184 } 17185 17186 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry buildPartial() { 17187 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry(this); 17188 int from_bitField0_ = bitField0_; 17189 int to_bitField0_ = 0; 17190 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 17191 to_bitField0_ |= 0x00000001; 17192 } 17193 result.id_ = id_; 17194 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 17195 to_bitField0_ |= 0x00000002; 17196 } 17197 result.str_ = str_; 17198 result.bitField0_ = to_bitField0_; 17199 onBuilt(); 17200 return result; 17201 } 17202 17203 public Builder mergeFrom(com.google.protobuf.Message other) { 17204 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry) { 17205 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry)other); 17206 } else { 17207 super.mergeFrom(other); 17208 return this; 17209 } 17210 } 17211 17212 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry other) { 17213 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.getDefaultInstance()) return this; 17214 if (other.hasId()) { 17215 setId(other.getId()); 17216 } 17217 if (other.hasStr()) { 17218 bitField0_ |= 0x00000002; 17219 str_ = other.str_; 17220 onChanged(); 17221 } 17222 this.mergeUnknownFields(other.getUnknownFields()); 17223 return this; 17224 } 17225 17226 public final boolean isInitialized() { 17227 return true; 17228 } 17229 17230 public Builder mergeFrom( 17231 com.google.protobuf.CodedInputStream input, 17232 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17233 throws java.io.IOException { 17234 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parsedMessage = null; 17235 try { 17236 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 17237 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17238 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry) e.getUnfinishedMessage(); 17239 throw e; 17240 } finally { 17241 if (parsedMessage != null) { 17242 mergeFrom(parsedMessage); 17243 } 17244 } 17245 return this; 17246 } 17247 private int bitField0_; 17248 17249 // optional uint32 id = 1; 17250 private int id_ ; 17251 /** 17252 * <code>optional uint32 id = 1;</code> 17253 */ 17254 public boolean hasId() { 17255 return ((bitField0_ & 0x00000001) == 0x00000001); 17256 } 17257 /** 17258 * <code>optional uint32 id = 1;</code> 17259 */ 17260 public int getId() { 17261 return id_; 17262 } 17263 /** 17264 * <code>optional uint32 id = 1;</code> 17265 */ 17266 public Builder setId(int value) { 17267 bitField0_ |= 0x00000001; 17268 id_ = value; 17269 onChanged(); 17270 return this; 17271 } 17272 /** 17273 * <code>optional uint32 id = 1;</code> 17274 */ 17275 public Builder clearId() { 17276 bitField0_ = (bitField0_ & ~0x00000001); 17277 id_ = 0; 17278 onChanged(); 17279 return this; 17280 } 17281 17282 // optional string str = 2; 17283 private java.lang.Object str_ = ""; 17284 /** 17285 * <code>optional string str = 2;</code> 17286 */ 17287 public boolean hasStr() { 17288 return ((bitField0_ & 0x00000002) == 0x00000002); 17289 } 17290 /** 17291 * <code>optional string str = 2;</code> 17292 */ 17293 public java.lang.String getStr() { 17294 java.lang.Object ref = str_; 17295 if (!(ref instanceof java.lang.String)) { 17296 java.lang.String s = ((com.google.protobuf.ByteString) ref) 17297 .toStringUtf8(); 17298 str_ = s; 17299 return s; 17300 } else { 17301 return (java.lang.String) ref; 17302 } 17303 } 17304 /** 17305 * <code>optional string str = 2;</code> 17306 */ 17307 public com.google.protobuf.ByteString 17308 getStrBytes() { 17309 java.lang.Object ref = str_; 17310 if (ref instanceof String) { 17311 com.google.protobuf.ByteString b = 17312 com.google.protobuf.ByteString.copyFromUtf8( 17313 (java.lang.String) ref); 17314 str_ = b; 17315 return b; 17316 } else { 17317 return (com.google.protobuf.ByteString) ref; 17318 } 17319 } 17320 /** 17321 * <code>optional string str = 2;</code> 17322 */ 17323 public Builder setStr( 17324 java.lang.String value) { 17325 if (value == null) { 17326 throw new NullPointerException(); 17327 } 17328 bitField0_ |= 0x00000002; 17329 str_ = value; 17330 onChanged(); 17331 return this; 17332 } 17333 /** 17334 * <code>optional string str = 2;</code> 17335 */ 17336 public Builder clearStr() { 17337 bitField0_ = (bitField0_ & ~0x00000002); 17338 str_ = getDefaultInstance().getStr(); 17339 onChanged(); 17340 return this; 17341 } 17342 /** 17343 * <code>optional string str = 2;</code> 17344 */ 17345 public Builder setStrBytes( 17346 com.google.protobuf.ByteString value) { 17347 if (value == null) { 17348 throw new NullPointerException(); 17349 } 17350 bitField0_ |= 0x00000002; 17351 str_ = value; 17352 onChanged(); 17353 return this; 17354 } 17355 17356 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.StringTableSection.Entry) 17357 } 17358 17359 static { 17360 defaultInstance = new Entry(true); 17361 defaultInstance.initFields(); 17362 } 17363 17364 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.StringTableSection.Entry) 17365 } 17366 17367 private int bitField0_; 17368 // optional uint32 numEntry = 1; 17369 public static final int NUMENTRY_FIELD_NUMBER = 1; 17370 private int numEntry_; 17371 /** 17372 * <code>optional uint32 numEntry = 1;</code> 17373 * 17374 * <pre> 17375 * repeated Entry 17376 * </pre> 17377 */ 17378 public boolean hasNumEntry() { 17379 return ((bitField0_ & 0x00000001) == 0x00000001); 17380 } 17381 /** 17382 * <code>optional uint32 numEntry = 1;</code> 17383 * 17384 * <pre> 17385 * repeated Entry 17386 * </pre> 17387 */ 17388 public int getNumEntry() { 17389 return numEntry_; 17390 } 17391 17392 private void initFields() { 17393 numEntry_ = 0; 17394 } 17395 private byte memoizedIsInitialized = -1; 17396 public final boolean isInitialized() { 17397 byte isInitialized = memoizedIsInitialized; 17398 if (isInitialized != -1) return isInitialized == 1; 17399 17400 memoizedIsInitialized = 1; 17401 return true; 17402 } 17403 17404 public void writeTo(com.google.protobuf.CodedOutputStream output) 17405 throws java.io.IOException { 17406 getSerializedSize(); 17407 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17408 output.writeUInt32(1, numEntry_); 17409 } 17410 getUnknownFields().writeTo(output); 17411 } 17412 17413 private int memoizedSerializedSize = -1; 17414 public int getSerializedSize() { 17415 int size = memoizedSerializedSize; 17416 if (size != -1) return size; 17417 17418 size = 0; 17419 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17420 size += com.google.protobuf.CodedOutputStream 17421 .computeUInt32Size(1, numEntry_); 17422 } 17423 size += getUnknownFields().getSerializedSize(); 17424 memoizedSerializedSize = size; 17425 return size; 17426 } 17427 17428 private static final long serialVersionUID = 0L; 17429 @java.lang.Override 17430 protected java.lang.Object writeReplace() 17431 throws java.io.ObjectStreamException { 17432 return super.writeReplace(); 17433 } 17434 17435 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17436 com.google.protobuf.ByteString data) 17437 throws com.google.protobuf.InvalidProtocolBufferException { 17438 return PARSER.parseFrom(data); 17439 } 17440 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17441 com.google.protobuf.ByteString data, 17442 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17443 throws com.google.protobuf.InvalidProtocolBufferException { 17444 return PARSER.parseFrom(data, extensionRegistry); 17445 } 17446 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom(byte[] data) 17447 throws com.google.protobuf.InvalidProtocolBufferException { 17448 return PARSER.parseFrom(data); 17449 } 17450 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17451 byte[] data, 17452 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17453 throws com.google.protobuf.InvalidProtocolBufferException { 17454 return PARSER.parseFrom(data, extensionRegistry); 17455 } 17456 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom(java.io.InputStream input) 17457 throws java.io.IOException { 17458 return PARSER.parseFrom(input); 17459 } 17460 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17461 java.io.InputStream input, 17462 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17463 throws java.io.IOException { 17464 return PARSER.parseFrom(input, extensionRegistry); 17465 } 17466 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseDelimitedFrom(java.io.InputStream input) 17467 throws java.io.IOException { 17468 return PARSER.parseDelimitedFrom(input); 17469 } 17470 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseDelimitedFrom( 17471 java.io.InputStream input, 17472 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17473 throws java.io.IOException { 17474 return PARSER.parseDelimitedFrom(input, extensionRegistry); 17475 } 17476 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17477 com.google.protobuf.CodedInputStream input) 17478 throws java.io.IOException { 17479 return PARSER.parseFrom(input); 17480 } 17481 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 17482 com.google.protobuf.CodedInputStream input, 17483 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17484 throws java.io.IOException { 17485 return PARSER.parseFrom(input, extensionRegistry); 17486 } 17487 17488 public static Builder newBuilder() { return Builder.create(); } 17489 public Builder newBuilderForType() { return newBuilder(); } 17490 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection prototype) { 17491 return newBuilder().mergeFrom(prototype); 17492 } 17493 public Builder toBuilder() { return newBuilder(this); } 17494 17495 @java.lang.Override 17496 protected Builder newBuilderForType( 17497 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17498 Builder builder = new Builder(parent); 17499 return builder; 17500 } 17501 /** 17502 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection} 17503 * 17504 * <pre> 17505 ** 17506 * This section maps string to id 17507 * NAME: STRING_TABLE 17508 * </pre> 17509 */ 17510 public static final class Builder extends 17511 com.google.protobuf.GeneratedMessage.Builder<Builder> 17512 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSectionOrBuilder { 17513 public static final com.google.protobuf.Descriptors.Descriptor 17514 getDescriptor() { 17515 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 17516 } 17517 17518 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17519 internalGetFieldAccessorTable() { 17520 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable 17521 .ensureFieldAccessorsInitialized( 17522 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Builder.class); 17523 } 17524 17525 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.newBuilder() 17526 private Builder() { 17527 maybeForceBuilderInitialization(); 17528 } 17529 17530 private Builder( 17531 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17532 super(parent); 17533 maybeForceBuilderInitialization(); 17534 } 17535 private void maybeForceBuilderInitialization() { 17536 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 17537 } 17538 } 17539 private static Builder create() { 17540 return new Builder(); 17541 } 17542 17543 public Builder clear() { 17544 super.clear(); 17545 numEntry_ = 0; 17546 bitField0_ = (bitField0_ & ~0x00000001); 17547 return this; 17548 } 17549 17550 public Builder clone() { 17551 return create().mergeFrom(buildPartial()); 17552 } 17553 17554 public com.google.protobuf.Descriptors.Descriptor 17555 getDescriptorForType() { 17556 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 17557 } 17558 17559 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection getDefaultInstanceForType() { 17560 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.getDefaultInstance(); 17561 } 17562 17563 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection build() { 17564 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection result = buildPartial(); 17565 if (!result.isInitialized()) { 17566 throw newUninitializedMessageException(result); 17567 } 17568 return result; 17569 } 17570 17571 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection buildPartial() { 17572 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection(this); 17573 int from_bitField0_ = bitField0_; 17574 int to_bitField0_ = 0; 17575 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 17576 to_bitField0_ |= 0x00000001; 17577 } 17578 result.numEntry_ = numEntry_; 17579 result.bitField0_ = to_bitField0_; 17580 onBuilt(); 17581 return result; 17582 } 17583 17584 public Builder mergeFrom(com.google.protobuf.Message other) { 17585 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection) { 17586 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection)other); 17587 } else { 17588 super.mergeFrom(other); 17589 return this; 17590 } 17591 } 17592 17593 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection other) { 17594 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.getDefaultInstance()) return this; 17595 if (other.hasNumEntry()) { 17596 setNumEntry(other.getNumEntry()); 17597 } 17598 this.mergeUnknownFields(other.getUnknownFields()); 17599 return this; 17600 } 17601 17602 public final boolean isInitialized() { 17603 return true; 17604 } 17605 17606 public Builder mergeFrom( 17607 com.google.protobuf.CodedInputStream input, 17608 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17609 throws java.io.IOException { 17610 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parsedMessage = null; 17611 try { 17612 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 17613 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17614 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection) e.getUnfinishedMessage(); 17615 throw e; 17616 } finally { 17617 if (parsedMessage != null) { 17618 mergeFrom(parsedMessage); 17619 } 17620 } 17621 return this; 17622 } 17623 private int bitField0_; 17624 17625 // optional uint32 numEntry = 1; 17626 private int numEntry_ ; 17627 /** 17628 * <code>optional uint32 numEntry = 1;</code> 17629 * 17630 * <pre> 17631 * repeated Entry 17632 * </pre> 17633 */ 17634 public boolean hasNumEntry() { 17635 return ((bitField0_ & 0x00000001) == 0x00000001); 17636 } 17637 /** 17638 * <code>optional uint32 numEntry = 1;</code> 17639 * 17640 * <pre> 17641 * repeated Entry 17642 * </pre> 17643 */ 17644 public int getNumEntry() { 17645 return numEntry_; 17646 } 17647 /** 17648 * <code>optional uint32 numEntry = 1;</code> 17649 * 17650 * <pre> 17651 * repeated Entry 17652 * </pre> 17653 */ 17654 public Builder setNumEntry(int value) { 17655 bitField0_ |= 0x00000001; 17656 numEntry_ = value; 17657 onChanged(); 17658 return this; 17659 } 17660 /** 17661 * <code>optional uint32 numEntry = 1;</code> 17662 * 17663 * <pre> 17664 * repeated Entry 17665 * </pre> 17666 */ 17667 public Builder clearNumEntry() { 17668 bitField0_ = (bitField0_ & ~0x00000001); 17669 numEntry_ = 0; 17670 onChanged(); 17671 return this; 17672 } 17673 17674 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.StringTableSection) 17675 } 17676 17677 static { 17678 defaultInstance = new StringTableSection(true); 17679 defaultInstance.initFields(); 17680 } 17681 17682 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.StringTableSection) 17683 } 17684 17685 public interface SecretManagerSectionOrBuilder 17686 extends com.google.protobuf.MessageOrBuilder { 17687 17688 // optional uint32 currentId = 1; 17689 /** 17690 * <code>optional uint32 currentId = 1;</code> 17691 */ 17692 boolean hasCurrentId(); 17693 /** 17694 * <code>optional uint32 currentId = 1;</code> 17695 */ 17696 int getCurrentId(); 17697 17698 // optional uint32 tokenSequenceNumber = 2; 17699 /** 17700 * <code>optional uint32 tokenSequenceNumber = 2;</code> 17701 */ 17702 boolean hasTokenSequenceNumber(); 17703 /** 17704 * <code>optional uint32 tokenSequenceNumber = 2;</code> 17705 */ 17706 int getTokenSequenceNumber(); 17707 17708 // optional uint32 numKeys = 3; 17709 /** 17710 * <code>optional uint32 numKeys = 3;</code> 17711 */ 17712 boolean hasNumKeys(); 17713 /** 17714 * <code>optional uint32 numKeys = 3;</code> 17715 */ 17716 int getNumKeys(); 17717 17718 // optional uint32 numTokens = 4; 17719 /** 17720 * <code>optional uint32 numTokens = 4;</code> 17721 * 17722 * <pre> 17723 * repeated DelegationKey keys 17724 * repeated PersistToken tokens 17725 * </pre> 17726 */ 17727 boolean hasNumTokens(); 17728 /** 17729 * <code>optional uint32 numTokens = 4;</code> 17730 * 17731 * <pre> 17732 * repeated DelegationKey keys 17733 * repeated PersistToken tokens 17734 * </pre> 17735 */ 17736 int getNumTokens(); 17737 } 17738 /** 17739 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection} 17740 */ 17741 public static final class SecretManagerSection extends 17742 com.google.protobuf.GeneratedMessage 17743 implements SecretManagerSectionOrBuilder { 17744 // Use SecretManagerSection.newBuilder() to construct. 17745 private SecretManagerSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 17746 super(builder); 17747 this.unknownFields = builder.getUnknownFields(); 17748 } 17749 private SecretManagerSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 17750 17751 private static final SecretManagerSection defaultInstance; 17752 public static SecretManagerSection getDefaultInstance() { 17753 return defaultInstance; 17754 } 17755 17756 public SecretManagerSection getDefaultInstanceForType() { 17757 return defaultInstance; 17758 } 17759 17760 private final com.google.protobuf.UnknownFieldSet unknownFields; 17761 @java.lang.Override 17762 public final com.google.protobuf.UnknownFieldSet 17763 getUnknownFields() { 17764 return this.unknownFields; 17765 } 17766 private SecretManagerSection( 17767 com.google.protobuf.CodedInputStream input, 17768 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17769 throws com.google.protobuf.InvalidProtocolBufferException { 17770 initFields(); 17771 int mutable_bitField0_ = 0; 17772 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 17773 com.google.protobuf.UnknownFieldSet.newBuilder(); 17774 try { 17775 boolean done = false; 17776 while (!done) { 17777 int tag = input.readTag(); 17778 switch (tag) { 17779 case 0: 17780 done = true; 17781 break; 17782 default: { 17783 if (!parseUnknownField(input, unknownFields, 17784 extensionRegistry, tag)) { 17785 done = true; 17786 } 17787 break; 17788 } 17789 case 8: { 17790 bitField0_ |= 0x00000001; 17791 currentId_ = input.readUInt32(); 17792 break; 17793 } 17794 case 16: { 17795 bitField0_ |= 0x00000002; 17796 tokenSequenceNumber_ = input.readUInt32(); 17797 break; 17798 } 17799 case 24: { 17800 bitField0_ |= 0x00000004; 17801 numKeys_ = input.readUInt32(); 17802 break; 17803 } 17804 case 32: { 17805 bitField0_ |= 0x00000008; 17806 numTokens_ = input.readUInt32(); 17807 break; 17808 } 17809 } 17810 } 17811 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17812 throw e.setUnfinishedMessage(this); 17813 } catch (java.io.IOException e) { 17814 throw new com.google.protobuf.InvalidProtocolBufferException( 17815 e.getMessage()).setUnfinishedMessage(this); 17816 } finally { 17817 this.unknownFields = unknownFields.build(); 17818 makeExtensionsImmutable(); 17819 } 17820 } 17821 public static final com.google.protobuf.Descriptors.Descriptor 17822 getDescriptor() { 17823 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 17824 } 17825 17826 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17827 internalGetFieldAccessorTable() { 17828 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable 17829 .ensureFieldAccessorsInitialized( 17830 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.Builder.class); 17831 } 17832 17833 public static com.google.protobuf.Parser<SecretManagerSection> PARSER = 17834 new com.google.protobuf.AbstractParser<SecretManagerSection>() { 17835 public SecretManagerSection parsePartialFrom( 17836 com.google.protobuf.CodedInputStream input, 17837 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17838 throws com.google.protobuf.InvalidProtocolBufferException { 17839 return new SecretManagerSection(input, extensionRegistry); 17840 } 17841 }; 17842 17843 @java.lang.Override 17844 public com.google.protobuf.Parser<SecretManagerSection> getParserForType() { 17845 return PARSER; 17846 } 17847 17848 public interface DelegationKeyOrBuilder 17849 extends com.google.protobuf.MessageOrBuilder { 17850 17851 // optional uint32 id = 1; 17852 /** 17853 * <code>optional uint32 id = 1;</code> 17854 */ 17855 boolean hasId(); 17856 /** 17857 * <code>optional uint32 id = 1;</code> 17858 */ 17859 int getId(); 17860 17861 // optional uint64 expiryDate = 2; 17862 /** 17863 * <code>optional uint64 expiryDate = 2;</code> 17864 */ 17865 boolean hasExpiryDate(); 17866 /** 17867 * <code>optional uint64 expiryDate = 2;</code> 17868 */ 17869 long getExpiryDate(); 17870 17871 // optional bytes key = 3; 17872 /** 17873 * <code>optional bytes key = 3;</code> 17874 */ 17875 boolean hasKey(); 17876 /** 17877 * <code>optional bytes key = 3;</code> 17878 */ 17879 com.google.protobuf.ByteString getKey(); 17880 } 17881 /** 17882 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey} 17883 */ 17884 public static final class DelegationKey extends 17885 com.google.protobuf.GeneratedMessage 17886 implements DelegationKeyOrBuilder { 17887 // Use DelegationKey.newBuilder() to construct. 17888 private DelegationKey(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 17889 super(builder); 17890 this.unknownFields = builder.getUnknownFields(); 17891 } 17892 private DelegationKey(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 17893 17894 private static final DelegationKey defaultInstance; 17895 public static DelegationKey getDefaultInstance() { 17896 return defaultInstance; 17897 } 17898 17899 public DelegationKey getDefaultInstanceForType() { 17900 return defaultInstance; 17901 } 17902 17903 private final com.google.protobuf.UnknownFieldSet unknownFields; 17904 @java.lang.Override 17905 public final com.google.protobuf.UnknownFieldSet 17906 getUnknownFields() { 17907 return this.unknownFields; 17908 } 17909 private DelegationKey( 17910 com.google.protobuf.CodedInputStream input, 17911 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17912 throws com.google.protobuf.InvalidProtocolBufferException { 17913 initFields(); 17914 int mutable_bitField0_ = 0; 17915 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 17916 com.google.protobuf.UnknownFieldSet.newBuilder(); 17917 try { 17918 boolean done = false; 17919 while (!done) { 17920 int tag = input.readTag(); 17921 switch (tag) { 17922 case 0: 17923 done = true; 17924 break; 17925 default: { 17926 if (!parseUnknownField(input, unknownFields, 17927 extensionRegistry, tag)) { 17928 done = true; 17929 } 17930 break; 17931 } 17932 case 8: { 17933 bitField0_ |= 0x00000001; 17934 id_ = input.readUInt32(); 17935 break; 17936 } 17937 case 16: { 17938 bitField0_ |= 0x00000002; 17939 expiryDate_ = input.readUInt64(); 17940 break; 17941 } 17942 case 26: { 17943 bitField0_ |= 0x00000004; 17944 key_ = input.readBytes(); 17945 break; 17946 } 17947 } 17948 } 17949 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17950 throw e.setUnfinishedMessage(this); 17951 } catch (java.io.IOException e) { 17952 throw new com.google.protobuf.InvalidProtocolBufferException( 17953 e.getMessage()).setUnfinishedMessage(this); 17954 } finally { 17955 this.unknownFields = unknownFields.build(); 17956 makeExtensionsImmutable(); 17957 } 17958 } 17959 public static final com.google.protobuf.Descriptors.Descriptor 17960 getDescriptor() { 17961 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 17962 } 17963 17964 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17965 internalGetFieldAccessorTable() { 17966 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable 17967 .ensureFieldAccessorsInitialized( 17968 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.Builder.class); 17969 } 17970 17971 public static com.google.protobuf.Parser<DelegationKey> PARSER = 17972 new com.google.protobuf.AbstractParser<DelegationKey>() { 17973 public DelegationKey parsePartialFrom( 17974 com.google.protobuf.CodedInputStream input, 17975 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17976 throws com.google.protobuf.InvalidProtocolBufferException { 17977 return new DelegationKey(input, extensionRegistry); 17978 } 17979 }; 17980 17981 @java.lang.Override 17982 public com.google.protobuf.Parser<DelegationKey> getParserForType() { 17983 return PARSER; 17984 } 17985 17986 private int bitField0_; 17987 // optional uint32 id = 1; 17988 public static final int ID_FIELD_NUMBER = 1; 17989 private int id_; 17990 /** 17991 * <code>optional uint32 id = 1;</code> 17992 */ 17993 public boolean hasId() { 17994 return ((bitField0_ & 0x00000001) == 0x00000001); 17995 } 17996 /** 17997 * <code>optional uint32 id = 1;</code> 17998 */ 17999 public int getId() { 18000 return id_; 18001 } 18002 18003 // optional uint64 expiryDate = 2; 18004 public static final int EXPIRYDATE_FIELD_NUMBER = 2; 18005 private long expiryDate_; 18006 /** 18007 * <code>optional uint64 expiryDate = 2;</code> 18008 */ 18009 public boolean hasExpiryDate() { 18010 return ((bitField0_ & 0x00000002) == 0x00000002); 18011 } 18012 /** 18013 * <code>optional uint64 expiryDate = 2;</code> 18014 */ 18015 public long getExpiryDate() { 18016 return expiryDate_; 18017 } 18018 18019 // optional bytes key = 3; 18020 public static final int KEY_FIELD_NUMBER = 3; 18021 private com.google.protobuf.ByteString key_; 18022 /** 18023 * <code>optional bytes key = 3;</code> 18024 */ 18025 public boolean hasKey() { 18026 return ((bitField0_ & 0x00000004) == 0x00000004); 18027 } 18028 /** 18029 * <code>optional bytes key = 3;</code> 18030 */ 18031 public com.google.protobuf.ByteString getKey() { 18032 return key_; 18033 } 18034 18035 private void initFields() { 18036 id_ = 0; 18037 expiryDate_ = 0L; 18038 key_ = com.google.protobuf.ByteString.EMPTY; 18039 } 18040 private byte memoizedIsInitialized = -1; 18041 public final boolean isInitialized() { 18042 byte isInitialized = memoizedIsInitialized; 18043 if (isInitialized != -1) return isInitialized == 1; 18044 18045 memoizedIsInitialized = 1; 18046 return true; 18047 } 18048 18049 public void writeTo(com.google.protobuf.CodedOutputStream output) 18050 throws java.io.IOException { 18051 getSerializedSize(); 18052 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18053 output.writeUInt32(1, id_); 18054 } 18055 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18056 output.writeUInt64(2, expiryDate_); 18057 } 18058 if (((bitField0_ & 0x00000004) == 0x00000004)) { 18059 output.writeBytes(3, key_); 18060 } 18061 getUnknownFields().writeTo(output); 18062 } 18063 18064 private int memoizedSerializedSize = -1; 18065 public int getSerializedSize() { 18066 int size = memoizedSerializedSize; 18067 if (size != -1) return size; 18068 18069 size = 0; 18070 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18071 size += com.google.protobuf.CodedOutputStream 18072 .computeUInt32Size(1, id_); 18073 } 18074 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18075 size += com.google.protobuf.CodedOutputStream 18076 .computeUInt64Size(2, expiryDate_); 18077 } 18078 if (((bitField0_ & 0x00000004) == 0x00000004)) { 18079 size += com.google.protobuf.CodedOutputStream 18080 .computeBytesSize(3, key_); 18081 } 18082 size += getUnknownFields().getSerializedSize(); 18083 memoizedSerializedSize = size; 18084 return size; 18085 } 18086 18087 private static final long serialVersionUID = 0L; 18088 @java.lang.Override 18089 protected java.lang.Object writeReplace() 18090 throws java.io.ObjectStreamException { 18091 return super.writeReplace(); 18092 } 18093 18094 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18095 com.google.protobuf.ByteString data) 18096 throws com.google.protobuf.InvalidProtocolBufferException { 18097 return PARSER.parseFrom(data); 18098 } 18099 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18100 com.google.protobuf.ByteString data, 18101 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18102 throws com.google.protobuf.InvalidProtocolBufferException { 18103 return PARSER.parseFrom(data, extensionRegistry); 18104 } 18105 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom(byte[] data) 18106 throws com.google.protobuf.InvalidProtocolBufferException { 18107 return PARSER.parseFrom(data); 18108 } 18109 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18110 byte[] data, 18111 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18112 throws com.google.protobuf.InvalidProtocolBufferException { 18113 return PARSER.parseFrom(data, extensionRegistry); 18114 } 18115 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom(java.io.InputStream input) 18116 throws java.io.IOException { 18117 return PARSER.parseFrom(input); 18118 } 18119 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18120 java.io.InputStream input, 18121 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18122 throws java.io.IOException { 18123 return PARSER.parseFrom(input, extensionRegistry); 18124 } 18125 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseDelimitedFrom(java.io.InputStream input) 18126 throws java.io.IOException { 18127 return PARSER.parseDelimitedFrom(input); 18128 } 18129 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseDelimitedFrom( 18130 java.io.InputStream input, 18131 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18132 throws java.io.IOException { 18133 return PARSER.parseDelimitedFrom(input, extensionRegistry); 18134 } 18135 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18136 com.google.protobuf.CodedInputStream input) 18137 throws java.io.IOException { 18138 return PARSER.parseFrom(input); 18139 } 18140 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 18141 com.google.protobuf.CodedInputStream input, 18142 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18143 throws java.io.IOException { 18144 return PARSER.parseFrom(input, extensionRegistry); 18145 } 18146 18147 public static Builder newBuilder() { return Builder.create(); } 18148 public Builder newBuilderForType() { return newBuilder(); } 18149 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey prototype) { 18150 return newBuilder().mergeFrom(prototype); 18151 } 18152 public Builder toBuilder() { return newBuilder(this); } 18153 18154 @java.lang.Override 18155 protected Builder newBuilderForType( 18156 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18157 Builder builder = new Builder(parent); 18158 return builder; 18159 } 18160 /** 18161 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey} 18162 */ 18163 public static final class Builder extends 18164 com.google.protobuf.GeneratedMessage.Builder<Builder> 18165 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKeyOrBuilder { 18166 public static final com.google.protobuf.Descriptors.Descriptor 18167 getDescriptor() { 18168 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 18169 } 18170 18171 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18172 internalGetFieldAccessorTable() { 18173 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable 18174 .ensureFieldAccessorsInitialized( 18175 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.Builder.class); 18176 } 18177 18178 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.newBuilder() 18179 private Builder() { 18180 maybeForceBuilderInitialization(); 18181 } 18182 18183 private Builder( 18184 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18185 super(parent); 18186 maybeForceBuilderInitialization(); 18187 } 18188 private void maybeForceBuilderInitialization() { 18189 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 18190 } 18191 } 18192 private static Builder create() { 18193 return new Builder(); 18194 } 18195 18196 public Builder clear() { 18197 super.clear(); 18198 id_ = 0; 18199 bitField0_ = (bitField0_ & ~0x00000001); 18200 expiryDate_ = 0L; 18201 bitField0_ = (bitField0_ & ~0x00000002); 18202 key_ = com.google.protobuf.ByteString.EMPTY; 18203 bitField0_ = (bitField0_ & ~0x00000004); 18204 return this; 18205 } 18206 18207 public Builder clone() { 18208 return create().mergeFrom(buildPartial()); 18209 } 18210 18211 public com.google.protobuf.Descriptors.Descriptor 18212 getDescriptorForType() { 18213 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 18214 } 18215 18216 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey getDefaultInstanceForType() { 18217 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.getDefaultInstance(); 18218 } 18219 18220 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey build() { 18221 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey result = buildPartial(); 18222 if (!result.isInitialized()) { 18223 throw newUninitializedMessageException(result); 18224 } 18225 return result; 18226 } 18227 18228 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey buildPartial() { 18229 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey(this); 18230 int from_bitField0_ = bitField0_; 18231 int to_bitField0_ = 0; 18232 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 18233 to_bitField0_ |= 0x00000001; 18234 } 18235 result.id_ = id_; 18236 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 18237 to_bitField0_ |= 0x00000002; 18238 } 18239 result.expiryDate_ = expiryDate_; 18240 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 18241 to_bitField0_ |= 0x00000004; 18242 } 18243 result.key_ = key_; 18244 result.bitField0_ = to_bitField0_; 18245 onBuilt(); 18246 return result; 18247 } 18248 18249 public Builder mergeFrom(com.google.protobuf.Message other) { 18250 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey) { 18251 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey)other); 18252 } else { 18253 super.mergeFrom(other); 18254 return this; 18255 } 18256 } 18257 18258 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey other) { 18259 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.getDefaultInstance()) return this; 18260 if (other.hasId()) { 18261 setId(other.getId()); 18262 } 18263 if (other.hasExpiryDate()) { 18264 setExpiryDate(other.getExpiryDate()); 18265 } 18266 if (other.hasKey()) { 18267 setKey(other.getKey()); 18268 } 18269 this.mergeUnknownFields(other.getUnknownFields()); 18270 return this; 18271 } 18272 18273 public final boolean isInitialized() { 18274 return true; 18275 } 18276 18277 public Builder mergeFrom( 18278 com.google.protobuf.CodedInputStream input, 18279 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18280 throws java.io.IOException { 18281 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parsedMessage = null; 18282 try { 18283 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 18284 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18285 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey) e.getUnfinishedMessage(); 18286 throw e; 18287 } finally { 18288 if (parsedMessage != null) { 18289 mergeFrom(parsedMessage); 18290 } 18291 } 18292 return this; 18293 } 18294 private int bitField0_; 18295 18296 // optional uint32 id = 1; 18297 private int id_ ; 18298 /** 18299 * <code>optional uint32 id = 1;</code> 18300 */ 18301 public boolean hasId() { 18302 return ((bitField0_ & 0x00000001) == 0x00000001); 18303 } 18304 /** 18305 * <code>optional uint32 id = 1;</code> 18306 */ 18307 public int getId() { 18308 return id_; 18309 } 18310 /** 18311 * <code>optional uint32 id = 1;</code> 18312 */ 18313 public Builder setId(int value) { 18314 bitField0_ |= 0x00000001; 18315 id_ = value; 18316 onChanged(); 18317 return this; 18318 } 18319 /** 18320 * <code>optional uint32 id = 1;</code> 18321 */ 18322 public Builder clearId() { 18323 bitField0_ = (bitField0_ & ~0x00000001); 18324 id_ = 0; 18325 onChanged(); 18326 return this; 18327 } 18328 18329 // optional uint64 expiryDate = 2; 18330 private long expiryDate_ ; 18331 /** 18332 * <code>optional uint64 expiryDate = 2;</code> 18333 */ 18334 public boolean hasExpiryDate() { 18335 return ((bitField0_ & 0x00000002) == 0x00000002); 18336 } 18337 /** 18338 * <code>optional uint64 expiryDate = 2;</code> 18339 */ 18340 public long getExpiryDate() { 18341 return expiryDate_; 18342 } 18343 /** 18344 * <code>optional uint64 expiryDate = 2;</code> 18345 */ 18346 public Builder setExpiryDate(long value) { 18347 bitField0_ |= 0x00000002; 18348 expiryDate_ = value; 18349 onChanged(); 18350 return this; 18351 } 18352 /** 18353 * <code>optional uint64 expiryDate = 2;</code> 18354 */ 18355 public Builder clearExpiryDate() { 18356 bitField0_ = (bitField0_ & ~0x00000002); 18357 expiryDate_ = 0L; 18358 onChanged(); 18359 return this; 18360 } 18361 18362 // optional bytes key = 3; 18363 private com.google.protobuf.ByteString key_ = com.google.protobuf.ByteString.EMPTY; 18364 /** 18365 * <code>optional bytes key = 3;</code> 18366 */ 18367 public boolean hasKey() { 18368 return ((bitField0_ & 0x00000004) == 0x00000004); 18369 } 18370 /** 18371 * <code>optional bytes key = 3;</code> 18372 */ 18373 public com.google.protobuf.ByteString getKey() { 18374 return key_; 18375 } 18376 /** 18377 * <code>optional bytes key = 3;</code> 18378 */ 18379 public Builder setKey(com.google.protobuf.ByteString value) { 18380 if (value == null) { 18381 throw new NullPointerException(); 18382 } 18383 bitField0_ |= 0x00000004; 18384 key_ = value; 18385 onChanged(); 18386 return this; 18387 } 18388 /** 18389 * <code>optional bytes key = 3;</code> 18390 */ 18391 public Builder clearKey() { 18392 bitField0_ = (bitField0_ & ~0x00000004); 18393 key_ = getDefaultInstance().getKey(); 18394 onChanged(); 18395 return this; 18396 } 18397 18398 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey) 18399 } 18400 18401 static { 18402 defaultInstance = new DelegationKey(true); 18403 defaultInstance.initFields(); 18404 } 18405 18406 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey) 18407 } 18408 18409 public interface PersistTokenOrBuilder 18410 extends com.google.protobuf.MessageOrBuilder { 18411 18412 // optional uint32 version = 1; 18413 /** 18414 * <code>optional uint32 version = 1;</code> 18415 */ 18416 boolean hasVersion(); 18417 /** 18418 * <code>optional uint32 version = 1;</code> 18419 */ 18420 int getVersion(); 18421 18422 // optional string owner = 2; 18423 /** 18424 * <code>optional string owner = 2;</code> 18425 */ 18426 boolean hasOwner(); 18427 /** 18428 * <code>optional string owner = 2;</code> 18429 */ 18430 java.lang.String getOwner(); 18431 /** 18432 * <code>optional string owner = 2;</code> 18433 */ 18434 com.google.protobuf.ByteString 18435 getOwnerBytes(); 18436 18437 // optional string renewer = 3; 18438 /** 18439 * <code>optional string renewer = 3;</code> 18440 */ 18441 boolean hasRenewer(); 18442 /** 18443 * <code>optional string renewer = 3;</code> 18444 */ 18445 java.lang.String getRenewer(); 18446 /** 18447 * <code>optional string renewer = 3;</code> 18448 */ 18449 com.google.protobuf.ByteString 18450 getRenewerBytes(); 18451 18452 // optional string realUser = 4; 18453 /** 18454 * <code>optional string realUser = 4;</code> 18455 */ 18456 boolean hasRealUser(); 18457 /** 18458 * <code>optional string realUser = 4;</code> 18459 */ 18460 java.lang.String getRealUser(); 18461 /** 18462 * <code>optional string realUser = 4;</code> 18463 */ 18464 com.google.protobuf.ByteString 18465 getRealUserBytes(); 18466 18467 // optional uint64 issueDate = 5; 18468 /** 18469 * <code>optional uint64 issueDate = 5;</code> 18470 */ 18471 boolean hasIssueDate(); 18472 /** 18473 * <code>optional uint64 issueDate = 5;</code> 18474 */ 18475 long getIssueDate(); 18476 18477 // optional uint64 maxDate = 6; 18478 /** 18479 * <code>optional uint64 maxDate = 6;</code> 18480 */ 18481 boolean hasMaxDate(); 18482 /** 18483 * <code>optional uint64 maxDate = 6;</code> 18484 */ 18485 long getMaxDate(); 18486 18487 // optional uint32 sequenceNumber = 7; 18488 /** 18489 * <code>optional uint32 sequenceNumber = 7;</code> 18490 */ 18491 boolean hasSequenceNumber(); 18492 /** 18493 * <code>optional uint32 sequenceNumber = 7;</code> 18494 */ 18495 int getSequenceNumber(); 18496 18497 // optional uint32 masterKeyId = 8; 18498 /** 18499 * <code>optional uint32 masterKeyId = 8;</code> 18500 */ 18501 boolean hasMasterKeyId(); 18502 /** 18503 * <code>optional uint32 masterKeyId = 8;</code> 18504 */ 18505 int getMasterKeyId(); 18506 18507 // optional uint64 expiryDate = 9; 18508 /** 18509 * <code>optional uint64 expiryDate = 9;</code> 18510 */ 18511 boolean hasExpiryDate(); 18512 /** 18513 * <code>optional uint64 expiryDate = 9;</code> 18514 */ 18515 long getExpiryDate(); 18516 } 18517 /** 18518 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.PersistToken} 18519 */ 18520 public static final class PersistToken extends 18521 com.google.protobuf.GeneratedMessage 18522 implements PersistTokenOrBuilder { 18523 // Use PersistToken.newBuilder() to construct. 18524 private PersistToken(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 18525 super(builder); 18526 this.unknownFields = builder.getUnknownFields(); 18527 } 18528 private PersistToken(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 18529 18530 private static final PersistToken defaultInstance; 18531 public static PersistToken getDefaultInstance() { 18532 return defaultInstance; 18533 } 18534 18535 public PersistToken getDefaultInstanceForType() { 18536 return defaultInstance; 18537 } 18538 18539 private final com.google.protobuf.UnknownFieldSet unknownFields; 18540 @java.lang.Override 18541 public final com.google.protobuf.UnknownFieldSet 18542 getUnknownFields() { 18543 return this.unknownFields; 18544 } 18545 private PersistToken( 18546 com.google.protobuf.CodedInputStream input, 18547 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18548 throws com.google.protobuf.InvalidProtocolBufferException { 18549 initFields(); 18550 int mutable_bitField0_ = 0; 18551 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 18552 com.google.protobuf.UnknownFieldSet.newBuilder(); 18553 try { 18554 boolean done = false; 18555 while (!done) { 18556 int tag = input.readTag(); 18557 switch (tag) { 18558 case 0: 18559 done = true; 18560 break; 18561 default: { 18562 if (!parseUnknownField(input, unknownFields, 18563 extensionRegistry, tag)) { 18564 done = true; 18565 } 18566 break; 18567 } 18568 case 8: { 18569 bitField0_ |= 0x00000001; 18570 version_ = input.readUInt32(); 18571 break; 18572 } 18573 case 18: { 18574 bitField0_ |= 0x00000002; 18575 owner_ = input.readBytes(); 18576 break; 18577 } 18578 case 26: { 18579 bitField0_ |= 0x00000004; 18580 renewer_ = input.readBytes(); 18581 break; 18582 } 18583 case 34: { 18584 bitField0_ |= 0x00000008; 18585 realUser_ = input.readBytes(); 18586 break; 18587 } 18588 case 40: { 18589 bitField0_ |= 0x00000010; 18590 issueDate_ = input.readUInt64(); 18591 break; 18592 } 18593 case 48: { 18594 bitField0_ |= 0x00000020; 18595 maxDate_ = input.readUInt64(); 18596 break; 18597 } 18598 case 56: { 18599 bitField0_ |= 0x00000040; 18600 sequenceNumber_ = input.readUInt32(); 18601 break; 18602 } 18603 case 64: { 18604 bitField0_ |= 0x00000080; 18605 masterKeyId_ = input.readUInt32(); 18606 break; 18607 } 18608 case 72: { 18609 bitField0_ |= 0x00000100; 18610 expiryDate_ = input.readUInt64(); 18611 break; 18612 } 18613 } 18614 } 18615 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18616 throw e.setUnfinishedMessage(this); 18617 } catch (java.io.IOException e) { 18618 throw new com.google.protobuf.InvalidProtocolBufferException( 18619 e.getMessage()).setUnfinishedMessage(this); 18620 } finally { 18621 this.unknownFields = unknownFields.build(); 18622 makeExtensionsImmutable(); 18623 } 18624 } 18625 public static final com.google.protobuf.Descriptors.Descriptor 18626 getDescriptor() { 18627 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 18628 } 18629 18630 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18631 internalGetFieldAccessorTable() { 18632 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable 18633 .ensureFieldAccessorsInitialized( 18634 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.Builder.class); 18635 } 18636 18637 public static com.google.protobuf.Parser<PersistToken> PARSER = 18638 new com.google.protobuf.AbstractParser<PersistToken>() { 18639 public PersistToken parsePartialFrom( 18640 com.google.protobuf.CodedInputStream input, 18641 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18642 throws com.google.protobuf.InvalidProtocolBufferException { 18643 return new PersistToken(input, extensionRegistry); 18644 } 18645 }; 18646 18647 @java.lang.Override 18648 public com.google.protobuf.Parser<PersistToken> getParserForType() { 18649 return PARSER; 18650 } 18651 18652 private int bitField0_; 18653 // optional uint32 version = 1; 18654 public static final int VERSION_FIELD_NUMBER = 1; 18655 private int version_; 18656 /** 18657 * <code>optional uint32 version = 1;</code> 18658 */ 18659 public boolean hasVersion() { 18660 return ((bitField0_ & 0x00000001) == 0x00000001); 18661 } 18662 /** 18663 * <code>optional uint32 version = 1;</code> 18664 */ 18665 public int getVersion() { 18666 return version_; 18667 } 18668 18669 // optional string owner = 2; 18670 public static final int OWNER_FIELD_NUMBER = 2; 18671 private java.lang.Object owner_; 18672 /** 18673 * <code>optional string owner = 2;</code> 18674 */ 18675 public boolean hasOwner() { 18676 return ((bitField0_ & 0x00000002) == 0x00000002); 18677 } 18678 /** 18679 * <code>optional string owner = 2;</code> 18680 */ 18681 public java.lang.String getOwner() { 18682 java.lang.Object ref = owner_; 18683 if (ref instanceof java.lang.String) { 18684 return (java.lang.String) ref; 18685 } else { 18686 com.google.protobuf.ByteString bs = 18687 (com.google.protobuf.ByteString) ref; 18688 java.lang.String s = bs.toStringUtf8(); 18689 if (bs.isValidUtf8()) { 18690 owner_ = s; 18691 } 18692 return s; 18693 } 18694 } 18695 /** 18696 * <code>optional string owner = 2;</code> 18697 */ 18698 public com.google.protobuf.ByteString 18699 getOwnerBytes() { 18700 java.lang.Object ref = owner_; 18701 if (ref instanceof java.lang.String) { 18702 com.google.protobuf.ByteString b = 18703 com.google.protobuf.ByteString.copyFromUtf8( 18704 (java.lang.String) ref); 18705 owner_ = b; 18706 return b; 18707 } else { 18708 return (com.google.protobuf.ByteString) ref; 18709 } 18710 } 18711 18712 // optional string renewer = 3; 18713 public static final int RENEWER_FIELD_NUMBER = 3; 18714 private java.lang.Object renewer_; 18715 /** 18716 * <code>optional string renewer = 3;</code> 18717 */ 18718 public boolean hasRenewer() { 18719 return ((bitField0_ & 0x00000004) == 0x00000004); 18720 } 18721 /** 18722 * <code>optional string renewer = 3;</code> 18723 */ 18724 public java.lang.String getRenewer() { 18725 java.lang.Object ref = renewer_; 18726 if (ref instanceof java.lang.String) { 18727 return (java.lang.String) ref; 18728 } else { 18729 com.google.protobuf.ByteString bs = 18730 (com.google.protobuf.ByteString) ref; 18731 java.lang.String s = bs.toStringUtf8(); 18732 if (bs.isValidUtf8()) { 18733 renewer_ = s; 18734 } 18735 return s; 18736 } 18737 } 18738 /** 18739 * <code>optional string renewer = 3;</code> 18740 */ 18741 public com.google.protobuf.ByteString 18742 getRenewerBytes() { 18743 java.lang.Object ref = renewer_; 18744 if (ref instanceof java.lang.String) { 18745 com.google.protobuf.ByteString b = 18746 com.google.protobuf.ByteString.copyFromUtf8( 18747 (java.lang.String) ref); 18748 renewer_ = b; 18749 return b; 18750 } else { 18751 return (com.google.protobuf.ByteString) ref; 18752 } 18753 } 18754 18755 // optional string realUser = 4; 18756 public static final int REALUSER_FIELD_NUMBER = 4; 18757 private java.lang.Object realUser_; 18758 /** 18759 * <code>optional string realUser = 4;</code> 18760 */ 18761 public boolean hasRealUser() { 18762 return ((bitField0_ & 0x00000008) == 0x00000008); 18763 } 18764 /** 18765 * <code>optional string realUser = 4;</code> 18766 */ 18767 public java.lang.String getRealUser() { 18768 java.lang.Object ref = realUser_; 18769 if (ref instanceof java.lang.String) { 18770 return (java.lang.String) ref; 18771 } else { 18772 com.google.protobuf.ByteString bs = 18773 (com.google.protobuf.ByteString) ref; 18774 java.lang.String s = bs.toStringUtf8(); 18775 if (bs.isValidUtf8()) { 18776 realUser_ = s; 18777 } 18778 return s; 18779 } 18780 } 18781 /** 18782 * <code>optional string realUser = 4;</code> 18783 */ 18784 public com.google.protobuf.ByteString 18785 getRealUserBytes() { 18786 java.lang.Object ref = realUser_; 18787 if (ref instanceof java.lang.String) { 18788 com.google.protobuf.ByteString b = 18789 com.google.protobuf.ByteString.copyFromUtf8( 18790 (java.lang.String) ref); 18791 realUser_ = b; 18792 return b; 18793 } else { 18794 return (com.google.protobuf.ByteString) ref; 18795 } 18796 } 18797 18798 // optional uint64 issueDate = 5; 18799 public static final int ISSUEDATE_FIELD_NUMBER = 5; 18800 private long issueDate_; 18801 /** 18802 * <code>optional uint64 issueDate = 5;</code> 18803 */ 18804 public boolean hasIssueDate() { 18805 return ((bitField0_ & 0x00000010) == 0x00000010); 18806 } 18807 /** 18808 * <code>optional uint64 issueDate = 5;</code> 18809 */ 18810 public long getIssueDate() { 18811 return issueDate_; 18812 } 18813 18814 // optional uint64 maxDate = 6; 18815 public static final int MAXDATE_FIELD_NUMBER = 6; 18816 private long maxDate_; 18817 /** 18818 * <code>optional uint64 maxDate = 6;</code> 18819 */ 18820 public boolean hasMaxDate() { 18821 return ((bitField0_ & 0x00000020) == 0x00000020); 18822 } 18823 /** 18824 * <code>optional uint64 maxDate = 6;</code> 18825 */ 18826 public long getMaxDate() { 18827 return maxDate_; 18828 } 18829 18830 // optional uint32 sequenceNumber = 7; 18831 public static final int SEQUENCENUMBER_FIELD_NUMBER = 7; 18832 private int sequenceNumber_; 18833 /** 18834 * <code>optional uint32 sequenceNumber = 7;</code> 18835 */ 18836 public boolean hasSequenceNumber() { 18837 return ((bitField0_ & 0x00000040) == 0x00000040); 18838 } 18839 /** 18840 * <code>optional uint32 sequenceNumber = 7;</code> 18841 */ 18842 public int getSequenceNumber() { 18843 return sequenceNumber_; 18844 } 18845 18846 // optional uint32 masterKeyId = 8; 18847 public static final int MASTERKEYID_FIELD_NUMBER = 8; 18848 private int masterKeyId_; 18849 /** 18850 * <code>optional uint32 masterKeyId = 8;</code> 18851 */ 18852 public boolean hasMasterKeyId() { 18853 return ((bitField0_ & 0x00000080) == 0x00000080); 18854 } 18855 /** 18856 * <code>optional uint32 masterKeyId = 8;</code> 18857 */ 18858 public int getMasterKeyId() { 18859 return masterKeyId_; 18860 } 18861 18862 // optional uint64 expiryDate = 9; 18863 public static final int EXPIRYDATE_FIELD_NUMBER = 9; 18864 private long expiryDate_; 18865 /** 18866 * <code>optional uint64 expiryDate = 9;</code> 18867 */ 18868 public boolean hasExpiryDate() { 18869 return ((bitField0_ & 0x00000100) == 0x00000100); 18870 } 18871 /** 18872 * <code>optional uint64 expiryDate = 9;</code> 18873 */ 18874 public long getExpiryDate() { 18875 return expiryDate_; 18876 } 18877 18878 private void initFields() { 18879 version_ = 0; 18880 owner_ = ""; 18881 renewer_ = ""; 18882 realUser_ = ""; 18883 issueDate_ = 0L; 18884 maxDate_ = 0L; 18885 sequenceNumber_ = 0; 18886 masterKeyId_ = 0; 18887 expiryDate_ = 0L; 18888 } 18889 private byte memoizedIsInitialized = -1; 18890 public final boolean isInitialized() { 18891 byte isInitialized = memoizedIsInitialized; 18892 if (isInitialized != -1) return isInitialized == 1; 18893 18894 memoizedIsInitialized = 1; 18895 return true; 18896 } 18897 18898 public void writeTo(com.google.protobuf.CodedOutputStream output) 18899 throws java.io.IOException { 18900 getSerializedSize(); 18901 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18902 output.writeUInt32(1, version_); 18903 } 18904 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18905 output.writeBytes(2, getOwnerBytes()); 18906 } 18907 if (((bitField0_ & 0x00000004) == 0x00000004)) { 18908 output.writeBytes(3, getRenewerBytes()); 18909 } 18910 if (((bitField0_ & 0x00000008) == 0x00000008)) { 18911 output.writeBytes(4, getRealUserBytes()); 18912 } 18913 if (((bitField0_ & 0x00000010) == 0x00000010)) { 18914 output.writeUInt64(5, issueDate_); 18915 } 18916 if (((bitField0_ & 0x00000020) == 0x00000020)) { 18917 output.writeUInt64(6, maxDate_); 18918 } 18919 if (((bitField0_ & 0x00000040) == 0x00000040)) { 18920 output.writeUInt32(7, sequenceNumber_); 18921 } 18922 if (((bitField0_ & 0x00000080) == 0x00000080)) { 18923 output.writeUInt32(8, masterKeyId_); 18924 } 18925 if (((bitField0_ & 0x00000100) == 0x00000100)) { 18926 output.writeUInt64(9, expiryDate_); 18927 } 18928 getUnknownFields().writeTo(output); 18929 } 18930 18931 private int memoizedSerializedSize = -1; 18932 public int getSerializedSize() { 18933 int size = memoizedSerializedSize; 18934 if (size != -1) return size; 18935 18936 size = 0; 18937 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18938 size += com.google.protobuf.CodedOutputStream 18939 .computeUInt32Size(1, version_); 18940 } 18941 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18942 size += com.google.protobuf.CodedOutputStream 18943 .computeBytesSize(2, getOwnerBytes()); 18944 } 18945 if (((bitField0_ & 0x00000004) == 0x00000004)) { 18946 size += com.google.protobuf.CodedOutputStream 18947 .computeBytesSize(3, getRenewerBytes()); 18948 } 18949 if (((bitField0_ & 0x00000008) == 0x00000008)) { 18950 size += com.google.protobuf.CodedOutputStream 18951 .computeBytesSize(4, getRealUserBytes()); 18952 } 18953 if (((bitField0_ & 0x00000010) == 0x00000010)) { 18954 size += com.google.protobuf.CodedOutputStream 18955 .computeUInt64Size(5, issueDate_); 18956 } 18957 if (((bitField0_ & 0x00000020) == 0x00000020)) { 18958 size += com.google.protobuf.CodedOutputStream 18959 .computeUInt64Size(6, maxDate_); 18960 } 18961 if (((bitField0_ & 0x00000040) == 0x00000040)) { 18962 size += com.google.protobuf.CodedOutputStream 18963 .computeUInt32Size(7, sequenceNumber_); 18964 } 18965 if (((bitField0_ & 0x00000080) == 0x00000080)) { 18966 size += com.google.protobuf.CodedOutputStream 18967 .computeUInt32Size(8, masterKeyId_); 18968 } 18969 if (((bitField0_ & 0x00000100) == 0x00000100)) { 18970 size += com.google.protobuf.CodedOutputStream 18971 .computeUInt64Size(9, expiryDate_); 18972 } 18973 size += getUnknownFields().getSerializedSize(); 18974 memoizedSerializedSize = size; 18975 return size; 18976 } 18977 18978 private static final long serialVersionUID = 0L; 18979 @java.lang.Override 18980 protected java.lang.Object writeReplace() 18981 throws java.io.ObjectStreamException { 18982 return super.writeReplace(); 18983 } 18984 18985 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 18986 com.google.protobuf.ByteString data) 18987 throws com.google.protobuf.InvalidProtocolBufferException { 18988 return PARSER.parseFrom(data); 18989 } 18990 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 18991 com.google.protobuf.ByteString data, 18992 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18993 throws com.google.protobuf.InvalidProtocolBufferException { 18994 return PARSER.parseFrom(data, extensionRegistry); 18995 } 18996 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom(byte[] data) 18997 throws com.google.protobuf.InvalidProtocolBufferException { 18998 return PARSER.parseFrom(data); 18999 } 19000 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 19001 byte[] data, 19002 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19003 throws com.google.protobuf.InvalidProtocolBufferException { 19004 return PARSER.parseFrom(data, extensionRegistry); 19005 } 19006 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom(java.io.InputStream input) 19007 throws java.io.IOException { 19008 return PARSER.parseFrom(input); 19009 } 19010 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 19011 java.io.InputStream input, 19012 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19013 throws java.io.IOException { 19014 return PARSER.parseFrom(input, extensionRegistry); 19015 } 19016 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseDelimitedFrom(java.io.InputStream input) 19017 throws java.io.IOException { 19018 return PARSER.parseDelimitedFrom(input); 19019 } 19020 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseDelimitedFrom( 19021 java.io.InputStream input, 19022 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19023 throws java.io.IOException { 19024 return PARSER.parseDelimitedFrom(input, extensionRegistry); 19025 } 19026 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 19027 com.google.protobuf.CodedInputStream input) 19028 throws java.io.IOException { 19029 return PARSER.parseFrom(input); 19030 } 19031 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 19032 com.google.protobuf.CodedInputStream input, 19033 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19034 throws java.io.IOException { 19035 return PARSER.parseFrom(input, extensionRegistry); 19036 } 19037 19038 public static Builder newBuilder() { return Builder.create(); } 19039 public Builder newBuilderForType() { return newBuilder(); } 19040 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken prototype) { 19041 return newBuilder().mergeFrom(prototype); 19042 } 19043 public Builder toBuilder() { return newBuilder(this); } 19044 19045 @java.lang.Override 19046 protected Builder newBuilderForType( 19047 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19048 Builder builder = new Builder(parent); 19049 return builder; 19050 } 19051 /** 19052 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.PersistToken} 19053 */ 19054 public static final class Builder extends 19055 com.google.protobuf.GeneratedMessage.Builder<Builder> 19056 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistTokenOrBuilder { 19057 public static final com.google.protobuf.Descriptors.Descriptor 19058 getDescriptor() { 19059 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 19060 } 19061 19062 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19063 internalGetFieldAccessorTable() { 19064 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable 19065 .ensureFieldAccessorsInitialized( 19066 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.Builder.class); 19067 } 19068 19069 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.newBuilder() 19070 private Builder() { 19071 maybeForceBuilderInitialization(); 19072 } 19073 19074 private Builder( 19075 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19076 super(parent); 19077 maybeForceBuilderInitialization(); 19078 } 19079 private void maybeForceBuilderInitialization() { 19080 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 19081 } 19082 } 19083 private static Builder create() { 19084 return new Builder(); 19085 } 19086 19087 public Builder clear() { 19088 super.clear(); 19089 version_ = 0; 19090 bitField0_ = (bitField0_ & ~0x00000001); 19091 owner_ = ""; 19092 bitField0_ = (bitField0_ & ~0x00000002); 19093 renewer_ = ""; 19094 bitField0_ = (bitField0_ & ~0x00000004); 19095 realUser_ = ""; 19096 bitField0_ = (bitField0_ & ~0x00000008); 19097 issueDate_ = 0L; 19098 bitField0_ = (bitField0_ & ~0x00000010); 19099 maxDate_ = 0L; 19100 bitField0_ = (bitField0_ & ~0x00000020); 19101 sequenceNumber_ = 0; 19102 bitField0_ = (bitField0_ & ~0x00000040); 19103 masterKeyId_ = 0; 19104 bitField0_ = (bitField0_ & ~0x00000080); 19105 expiryDate_ = 0L; 19106 bitField0_ = (bitField0_ & ~0x00000100); 19107 return this; 19108 } 19109 19110 public Builder clone() { 19111 return create().mergeFrom(buildPartial()); 19112 } 19113 19114 public com.google.protobuf.Descriptors.Descriptor 19115 getDescriptorForType() { 19116 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 19117 } 19118 19119 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken getDefaultInstanceForType() { 19120 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.getDefaultInstance(); 19121 } 19122 19123 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken build() { 19124 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken result = buildPartial(); 19125 if (!result.isInitialized()) { 19126 throw newUninitializedMessageException(result); 19127 } 19128 return result; 19129 } 19130 19131 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken buildPartial() { 19132 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken(this); 19133 int from_bitField0_ = bitField0_; 19134 int to_bitField0_ = 0; 19135 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 19136 to_bitField0_ |= 0x00000001; 19137 } 19138 result.version_ = version_; 19139 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 19140 to_bitField0_ |= 0x00000002; 19141 } 19142 result.owner_ = owner_; 19143 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 19144 to_bitField0_ |= 0x00000004; 19145 } 19146 result.renewer_ = renewer_; 19147 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 19148 to_bitField0_ |= 0x00000008; 19149 } 19150 result.realUser_ = realUser_; 19151 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 19152 to_bitField0_ |= 0x00000010; 19153 } 19154 result.issueDate_ = issueDate_; 19155 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 19156 to_bitField0_ |= 0x00000020; 19157 } 19158 result.maxDate_ = maxDate_; 19159 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 19160 to_bitField0_ |= 0x00000040; 19161 } 19162 result.sequenceNumber_ = sequenceNumber_; 19163 if (((from_bitField0_ & 0x00000080) == 0x00000080)) { 19164 to_bitField0_ |= 0x00000080; 19165 } 19166 result.masterKeyId_ = masterKeyId_; 19167 if (((from_bitField0_ & 0x00000100) == 0x00000100)) { 19168 to_bitField0_ |= 0x00000100; 19169 } 19170 result.expiryDate_ = expiryDate_; 19171 result.bitField0_ = to_bitField0_; 19172 onBuilt(); 19173 return result; 19174 } 19175 19176 public Builder mergeFrom(com.google.protobuf.Message other) { 19177 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken) { 19178 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken)other); 19179 } else { 19180 super.mergeFrom(other); 19181 return this; 19182 } 19183 } 19184 19185 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken other) { 19186 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.getDefaultInstance()) return this; 19187 if (other.hasVersion()) { 19188 setVersion(other.getVersion()); 19189 } 19190 if (other.hasOwner()) { 19191 bitField0_ |= 0x00000002; 19192 owner_ = other.owner_; 19193 onChanged(); 19194 } 19195 if (other.hasRenewer()) { 19196 bitField0_ |= 0x00000004; 19197 renewer_ = other.renewer_; 19198 onChanged(); 19199 } 19200 if (other.hasRealUser()) { 19201 bitField0_ |= 0x00000008; 19202 realUser_ = other.realUser_; 19203 onChanged(); 19204 } 19205 if (other.hasIssueDate()) { 19206 setIssueDate(other.getIssueDate()); 19207 } 19208 if (other.hasMaxDate()) { 19209 setMaxDate(other.getMaxDate()); 19210 } 19211 if (other.hasSequenceNumber()) { 19212 setSequenceNumber(other.getSequenceNumber()); 19213 } 19214 if (other.hasMasterKeyId()) { 19215 setMasterKeyId(other.getMasterKeyId()); 19216 } 19217 if (other.hasExpiryDate()) { 19218 setExpiryDate(other.getExpiryDate()); 19219 } 19220 this.mergeUnknownFields(other.getUnknownFields()); 19221 return this; 19222 } 19223 19224 public final boolean isInitialized() { 19225 return true; 19226 } 19227 19228 public Builder mergeFrom( 19229 com.google.protobuf.CodedInputStream input, 19230 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19231 throws java.io.IOException { 19232 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parsedMessage = null; 19233 try { 19234 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 19235 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 19236 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken) e.getUnfinishedMessage(); 19237 throw e; 19238 } finally { 19239 if (parsedMessage != null) { 19240 mergeFrom(parsedMessage); 19241 } 19242 } 19243 return this; 19244 } 19245 private int bitField0_; 19246 19247 // optional uint32 version = 1; 19248 private int version_ ; 19249 /** 19250 * <code>optional uint32 version = 1;</code> 19251 */ 19252 public boolean hasVersion() { 19253 return ((bitField0_ & 0x00000001) == 0x00000001); 19254 } 19255 /** 19256 * <code>optional uint32 version = 1;</code> 19257 */ 19258 public int getVersion() { 19259 return version_; 19260 } 19261 /** 19262 * <code>optional uint32 version = 1;</code> 19263 */ 19264 public Builder setVersion(int value) { 19265 bitField0_ |= 0x00000001; 19266 version_ = value; 19267 onChanged(); 19268 return this; 19269 } 19270 /** 19271 * <code>optional uint32 version = 1;</code> 19272 */ 19273 public Builder clearVersion() { 19274 bitField0_ = (bitField0_ & ~0x00000001); 19275 version_ = 0; 19276 onChanged(); 19277 return this; 19278 } 19279 19280 // optional string owner = 2; 19281 private java.lang.Object owner_ = ""; 19282 /** 19283 * <code>optional string owner = 2;</code> 19284 */ 19285 public boolean hasOwner() { 19286 return ((bitField0_ & 0x00000002) == 0x00000002); 19287 } 19288 /** 19289 * <code>optional string owner = 2;</code> 19290 */ 19291 public java.lang.String getOwner() { 19292 java.lang.Object ref = owner_; 19293 if (!(ref instanceof java.lang.String)) { 19294 java.lang.String s = ((com.google.protobuf.ByteString) ref) 19295 .toStringUtf8(); 19296 owner_ = s; 19297 return s; 19298 } else { 19299 return (java.lang.String) ref; 19300 } 19301 } 19302 /** 19303 * <code>optional string owner = 2;</code> 19304 */ 19305 public com.google.protobuf.ByteString 19306 getOwnerBytes() { 19307 java.lang.Object ref = owner_; 19308 if (ref instanceof String) { 19309 com.google.protobuf.ByteString b = 19310 com.google.protobuf.ByteString.copyFromUtf8( 19311 (java.lang.String) ref); 19312 owner_ = b; 19313 return b; 19314 } else { 19315 return (com.google.protobuf.ByteString) ref; 19316 } 19317 } 19318 /** 19319 * <code>optional string owner = 2;</code> 19320 */ 19321 public Builder setOwner( 19322 java.lang.String value) { 19323 if (value == null) { 19324 throw new NullPointerException(); 19325 } 19326 bitField0_ |= 0x00000002; 19327 owner_ = value; 19328 onChanged(); 19329 return this; 19330 } 19331 /** 19332 * <code>optional string owner = 2;</code> 19333 */ 19334 public Builder clearOwner() { 19335 bitField0_ = (bitField0_ & ~0x00000002); 19336 owner_ = getDefaultInstance().getOwner(); 19337 onChanged(); 19338 return this; 19339 } 19340 /** 19341 * <code>optional string owner = 2;</code> 19342 */ 19343 public Builder setOwnerBytes( 19344 com.google.protobuf.ByteString value) { 19345 if (value == null) { 19346 throw new NullPointerException(); 19347 } 19348 bitField0_ |= 0x00000002; 19349 owner_ = value; 19350 onChanged(); 19351 return this; 19352 } 19353 19354 // optional string renewer = 3; 19355 private java.lang.Object renewer_ = ""; 19356 /** 19357 * <code>optional string renewer = 3;</code> 19358 */ 19359 public boolean hasRenewer() { 19360 return ((bitField0_ & 0x00000004) == 0x00000004); 19361 } 19362 /** 19363 * <code>optional string renewer = 3;</code> 19364 */ 19365 public java.lang.String getRenewer() { 19366 java.lang.Object ref = renewer_; 19367 if (!(ref instanceof java.lang.String)) { 19368 java.lang.String s = ((com.google.protobuf.ByteString) ref) 19369 .toStringUtf8(); 19370 renewer_ = s; 19371 return s; 19372 } else { 19373 return (java.lang.String) ref; 19374 } 19375 } 19376 /** 19377 * <code>optional string renewer = 3;</code> 19378 */ 19379 public com.google.protobuf.ByteString 19380 getRenewerBytes() { 19381 java.lang.Object ref = renewer_; 19382 if (ref instanceof String) { 19383 com.google.protobuf.ByteString b = 19384 com.google.protobuf.ByteString.copyFromUtf8( 19385 (java.lang.String) ref); 19386 renewer_ = b; 19387 return b; 19388 } else { 19389 return (com.google.protobuf.ByteString) ref; 19390 } 19391 } 19392 /** 19393 * <code>optional string renewer = 3;</code> 19394 */ 19395 public Builder setRenewer( 19396 java.lang.String value) { 19397 if (value == null) { 19398 throw new NullPointerException(); 19399 } 19400 bitField0_ |= 0x00000004; 19401 renewer_ = value; 19402 onChanged(); 19403 return this; 19404 } 19405 /** 19406 * <code>optional string renewer = 3;</code> 19407 */ 19408 public Builder clearRenewer() { 19409 bitField0_ = (bitField0_ & ~0x00000004); 19410 renewer_ = getDefaultInstance().getRenewer(); 19411 onChanged(); 19412 return this; 19413 } 19414 /** 19415 * <code>optional string renewer = 3;</code> 19416 */ 19417 public Builder setRenewerBytes( 19418 com.google.protobuf.ByteString value) { 19419 if (value == null) { 19420 throw new NullPointerException(); 19421 } 19422 bitField0_ |= 0x00000004; 19423 renewer_ = value; 19424 onChanged(); 19425 return this; 19426 } 19427 19428 // optional string realUser = 4; 19429 private java.lang.Object realUser_ = ""; 19430 /** 19431 * <code>optional string realUser = 4;</code> 19432 */ 19433 public boolean hasRealUser() { 19434 return ((bitField0_ & 0x00000008) == 0x00000008); 19435 } 19436 /** 19437 * <code>optional string realUser = 4;</code> 19438 */ 19439 public java.lang.String getRealUser() { 19440 java.lang.Object ref = realUser_; 19441 if (!(ref instanceof java.lang.String)) { 19442 java.lang.String s = ((com.google.protobuf.ByteString) ref) 19443 .toStringUtf8(); 19444 realUser_ = s; 19445 return s; 19446 } else { 19447 return (java.lang.String) ref; 19448 } 19449 } 19450 /** 19451 * <code>optional string realUser = 4;</code> 19452 */ 19453 public com.google.protobuf.ByteString 19454 getRealUserBytes() { 19455 java.lang.Object ref = realUser_; 19456 if (ref instanceof String) { 19457 com.google.protobuf.ByteString b = 19458 com.google.protobuf.ByteString.copyFromUtf8( 19459 (java.lang.String) ref); 19460 realUser_ = b; 19461 return b; 19462 } else { 19463 return (com.google.protobuf.ByteString) ref; 19464 } 19465 } 19466 /** 19467 * <code>optional string realUser = 4;</code> 19468 */ 19469 public Builder setRealUser( 19470 java.lang.String value) { 19471 if (value == null) { 19472 throw new NullPointerException(); 19473 } 19474 bitField0_ |= 0x00000008; 19475 realUser_ = value; 19476 onChanged(); 19477 return this; 19478 } 19479 /** 19480 * <code>optional string realUser = 4;</code> 19481 */ 19482 public Builder clearRealUser() { 19483 bitField0_ = (bitField0_ & ~0x00000008); 19484 realUser_ = getDefaultInstance().getRealUser(); 19485 onChanged(); 19486 return this; 19487 } 19488 /** 19489 * <code>optional string realUser = 4;</code> 19490 */ 19491 public Builder setRealUserBytes( 19492 com.google.protobuf.ByteString value) { 19493 if (value == null) { 19494 throw new NullPointerException(); 19495 } 19496 bitField0_ |= 0x00000008; 19497 realUser_ = value; 19498 onChanged(); 19499 return this; 19500 } 19501 19502 // optional uint64 issueDate = 5; 19503 private long issueDate_ ; 19504 /** 19505 * <code>optional uint64 issueDate = 5;</code> 19506 */ 19507 public boolean hasIssueDate() { 19508 return ((bitField0_ & 0x00000010) == 0x00000010); 19509 } 19510 /** 19511 * <code>optional uint64 issueDate = 5;</code> 19512 */ 19513 public long getIssueDate() { 19514 return issueDate_; 19515 } 19516 /** 19517 * <code>optional uint64 issueDate = 5;</code> 19518 */ 19519 public Builder setIssueDate(long value) { 19520 bitField0_ |= 0x00000010; 19521 issueDate_ = value; 19522 onChanged(); 19523 return this; 19524 } 19525 /** 19526 * <code>optional uint64 issueDate = 5;</code> 19527 */ 19528 public Builder clearIssueDate() { 19529 bitField0_ = (bitField0_ & ~0x00000010); 19530 issueDate_ = 0L; 19531 onChanged(); 19532 return this; 19533 } 19534 19535 // optional uint64 maxDate = 6; 19536 private long maxDate_ ; 19537 /** 19538 * <code>optional uint64 maxDate = 6;</code> 19539 */ 19540 public boolean hasMaxDate() { 19541 return ((bitField0_ & 0x00000020) == 0x00000020); 19542 } 19543 /** 19544 * <code>optional uint64 maxDate = 6;</code> 19545 */ 19546 public long getMaxDate() { 19547 return maxDate_; 19548 } 19549 /** 19550 * <code>optional uint64 maxDate = 6;</code> 19551 */ 19552 public Builder setMaxDate(long value) { 19553 bitField0_ |= 0x00000020; 19554 maxDate_ = value; 19555 onChanged(); 19556 return this; 19557 } 19558 /** 19559 * <code>optional uint64 maxDate = 6;</code> 19560 */ 19561 public Builder clearMaxDate() { 19562 bitField0_ = (bitField0_ & ~0x00000020); 19563 maxDate_ = 0L; 19564 onChanged(); 19565 return this; 19566 } 19567 19568 // optional uint32 sequenceNumber = 7; 19569 private int sequenceNumber_ ; 19570 /** 19571 * <code>optional uint32 sequenceNumber = 7;</code> 19572 */ 19573 public boolean hasSequenceNumber() { 19574 return ((bitField0_ & 0x00000040) == 0x00000040); 19575 } 19576 /** 19577 * <code>optional uint32 sequenceNumber = 7;</code> 19578 */ 19579 public int getSequenceNumber() { 19580 return sequenceNumber_; 19581 } 19582 /** 19583 * <code>optional uint32 sequenceNumber = 7;</code> 19584 */ 19585 public Builder setSequenceNumber(int value) { 19586 bitField0_ |= 0x00000040; 19587 sequenceNumber_ = value; 19588 onChanged(); 19589 return this; 19590 } 19591 /** 19592 * <code>optional uint32 sequenceNumber = 7;</code> 19593 */ 19594 public Builder clearSequenceNumber() { 19595 bitField0_ = (bitField0_ & ~0x00000040); 19596 sequenceNumber_ = 0; 19597 onChanged(); 19598 return this; 19599 } 19600 19601 // optional uint32 masterKeyId = 8; 19602 private int masterKeyId_ ; 19603 /** 19604 * <code>optional uint32 masterKeyId = 8;</code> 19605 */ 19606 public boolean hasMasterKeyId() { 19607 return ((bitField0_ & 0x00000080) == 0x00000080); 19608 } 19609 /** 19610 * <code>optional uint32 masterKeyId = 8;</code> 19611 */ 19612 public int getMasterKeyId() { 19613 return masterKeyId_; 19614 } 19615 /** 19616 * <code>optional uint32 masterKeyId = 8;</code> 19617 */ 19618 public Builder setMasterKeyId(int value) { 19619 bitField0_ |= 0x00000080; 19620 masterKeyId_ = value; 19621 onChanged(); 19622 return this; 19623 } 19624 /** 19625 * <code>optional uint32 masterKeyId = 8;</code> 19626 */ 19627 public Builder clearMasterKeyId() { 19628 bitField0_ = (bitField0_ & ~0x00000080); 19629 masterKeyId_ = 0; 19630 onChanged(); 19631 return this; 19632 } 19633 19634 // optional uint64 expiryDate = 9; 19635 private long expiryDate_ ; 19636 /** 19637 * <code>optional uint64 expiryDate = 9;</code> 19638 */ 19639 public boolean hasExpiryDate() { 19640 return ((bitField0_ & 0x00000100) == 0x00000100); 19641 } 19642 /** 19643 * <code>optional uint64 expiryDate = 9;</code> 19644 */ 19645 public long getExpiryDate() { 19646 return expiryDate_; 19647 } 19648 /** 19649 * <code>optional uint64 expiryDate = 9;</code> 19650 */ 19651 public Builder setExpiryDate(long value) { 19652 bitField0_ |= 0x00000100; 19653 expiryDate_ = value; 19654 onChanged(); 19655 return this; 19656 } 19657 /** 19658 * <code>optional uint64 expiryDate = 9;</code> 19659 */ 19660 public Builder clearExpiryDate() { 19661 bitField0_ = (bitField0_ & ~0x00000100); 19662 expiryDate_ = 0L; 19663 onChanged(); 19664 return this; 19665 } 19666 19667 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection.PersistToken) 19668 } 19669 19670 static { 19671 defaultInstance = new PersistToken(true); 19672 defaultInstance.initFields(); 19673 } 19674 19675 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection.PersistToken) 19676 } 19677 19678 private int bitField0_; 19679 // optional uint32 currentId = 1; 19680 public static final int CURRENTID_FIELD_NUMBER = 1; 19681 private int currentId_; 19682 /** 19683 * <code>optional uint32 currentId = 1;</code> 19684 */ 19685 public boolean hasCurrentId() { 19686 return ((bitField0_ & 0x00000001) == 0x00000001); 19687 } 19688 /** 19689 * <code>optional uint32 currentId = 1;</code> 19690 */ 19691 public int getCurrentId() { 19692 return currentId_; 19693 } 19694 19695 // optional uint32 tokenSequenceNumber = 2; 19696 public static final int TOKENSEQUENCENUMBER_FIELD_NUMBER = 2; 19697 private int tokenSequenceNumber_; 19698 /** 19699 * <code>optional uint32 tokenSequenceNumber = 2;</code> 19700 */ 19701 public boolean hasTokenSequenceNumber() { 19702 return ((bitField0_ & 0x00000002) == 0x00000002); 19703 } 19704 /** 19705 * <code>optional uint32 tokenSequenceNumber = 2;</code> 19706 */ 19707 public int getTokenSequenceNumber() { 19708 return tokenSequenceNumber_; 19709 } 19710 19711 // optional uint32 numKeys = 3; 19712 public static final int NUMKEYS_FIELD_NUMBER = 3; 19713 private int numKeys_; 19714 /** 19715 * <code>optional uint32 numKeys = 3;</code> 19716 */ 19717 public boolean hasNumKeys() { 19718 return ((bitField0_ & 0x00000004) == 0x00000004); 19719 } 19720 /** 19721 * <code>optional uint32 numKeys = 3;</code> 19722 */ 19723 public int getNumKeys() { 19724 return numKeys_; 19725 } 19726 19727 // optional uint32 numTokens = 4; 19728 public static final int NUMTOKENS_FIELD_NUMBER = 4; 19729 private int numTokens_; 19730 /** 19731 * <code>optional uint32 numTokens = 4;</code> 19732 * 19733 * <pre> 19734 * repeated DelegationKey keys 19735 * repeated PersistToken tokens 19736 * </pre> 19737 */ 19738 public boolean hasNumTokens() { 19739 return ((bitField0_ & 0x00000008) == 0x00000008); 19740 } 19741 /** 19742 * <code>optional uint32 numTokens = 4;</code> 19743 * 19744 * <pre> 19745 * repeated DelegationKey keys 19746 * repeated PersistToken tokens 19747 * </pre> 19748 */ 19749 public int getNumTokens() { 19750 return numTokens_; 19751 } 19752 19753 private void initFields() { 19754 currentId_ = 0; 19755 tokenSequenceNumber_ = 0; 19756 numKeys_ = 0; 19757 numTokens_ = 0; 19758 } 19759 private byte memoizedIsInitialized = -1; 19760 public final boolean isInitialized() { 19761 byte isInitialized = memoizedIsInitialized; 19762 if (isInitialized != -1) return isInitialized == 1; 19763 19764 memoizedIsInitialized = 1; 19765 return true; 19766 } 19767 19768 public void writeTo(com.google.protobuf.CodedOutputStream output) 19769 throws java.io.IOException { 19770 getSerializedSize(); 19771 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19772 output.writeUInt32(1, currentId_); 19773 } 19774 if (((bitField0_ & 0x00000002) == 0x00000002)) { 19775 output.writeUInt32(2, tokenSequenceNumber_); 19776 } 19777 if (((bitField0_ & 0x00000004) == 0x00000004)) { 19778 output.writeUInt32(3, numKeys_); 19779 } 19780 if (((bitField0_ & 0x00000008) == 0x00000008)) { 19781 output.writeUInt32(4, numTokens_); 19782 } 19783 getUnknownFields().writeTo(output); 19784 } 19785 19786 private int memoizedSerializedSize = -1; 19787 public int getSerializedSize() { 19788 int size = memoizedSerializedSize; 19789 if (size != -1) return size; 19790 19791 size = 0; 19792 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19793 size += com.google.protobuf.CodedOutputStream 19794 .computeUInt32Size(1, currentId_); 19795 } 19796 if (((bitField0_ & 0x00000002) == 0x00000002)) { 19797 size += com.google.protobuf.CodedOutputStream 19798 .computeUInt32Size(2, tokenSequenceNumber_); 19799 } 19800 if (((bitField0_ & 0x00000004) == 0x00000004)) { 19801 size += com.google.protobuf.CodedOutputStream 19802 .computeUInt32Size(3, numKeys_); 19803 } 19804 if (((bitField0_ & 0x00000008) == 0x00000008)) { 19805 size += com.google.protobuf.CodedOutputStream 19806 .computeUInt32Size(4, numTokens_); 19807 } 19808 size += getUnknownFields().getSerializedSize(); 19809 memoizedSerializedSize = size; 19810 return size; 19811 } 19812 19813 private static final long serialVersionUID = 0L; 19814 @java.lang.Override 19815 protected java.lang.Object writeReplace() 19816 throws java.io.ObjectStreamException { 19817 return super.writeReplace(); 19818 } 19819 19820 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19821 com.google.protobuf.ByteString data) 19822 throws com.google.protobuf.InvalidProtocolBufferException { 19823 return PARSER.parseFrom(data); 19824 } 19825 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19826 com.google.protobuf.ByteString data, 19827 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19828 throws com.google.protobuf.InvalidProtocolBufferException { 19829 return PARSER.parseFrom(data, extensionRegistry); 19830 } 19831 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom(byte[] data) 19832 throws com.google.protobuf.InvalidProtocolBufferException { 19833 return PARSER.parseFrom(data); 19834 } 19835 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19836 byte[] data, 19837 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19838 throws com.google.protobuf.InvalidProtocolBufferException { 19839 return PARSER.parseFrom(data, extensionRegistry); 19840 } 19841 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom(java.io.InputStream input) 19842 throws java.io.IOException { 19843 return PARSER.parseFrom(input); 19844 } 19845 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19846 java.io.InputStream input, 19847 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19848 throws java.io.IOException { 19849 return PARSER.parseFrom(input, extensionRegistry); 19850 } 19851 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseDelimitedFrom(java.io.InputStream input) 19852 throws java.io.IOException { 19853 return PARSER.parseDelimitedFrom(input); 19854 } 19855 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseDelimitedFrom( 19856 java.io.InputStream input, 19857 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19858 throws java.io.IOException { 19859 return PARSER.parseDelimitedFrom(input, extensionRegistry); 19860 } 19861 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19862 com.google.protobuf.CodedInputStream input) 19863 throws java.io.IOException { 19864 return PARSER.parseFrom(input); 19865 } 19866 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 19867 com.google.protobuf.CodedInputStream input, 19868 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19869 throws java.io.IOException { 19870 return PARSER.parseFrom(input, extensionRegistry); 19871 } 19872 19873 public static Builder newBuilder() { return Builder.create(); } 19874 public Builder newBuilderForType() { return newBuilder(); } 19875 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection prototype) { 19876 return newBuilder().mergeFrom(prototype); 19877 } 19878 public Builder toBuilder() { return newBuilder(this); } 19879 19880 @java.lang.Override 19881 protected Builder newBuilderForType( 19882 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19883 Builder builder = new Builder(parent); 19884 return builder; 19885 } 19886 /** 19887 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection} 19888 */ 19889 public static final class Builder extends 19890 com.google.protobuf.GeneratedMessage.Builder<Builder> 19891 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSectionOrBuilder { 19892 public static final com.google.protobuf.Descriptors.Descriptor 19893 getDescriptor() { 19894 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 19895 } 19896 19897 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19898 internalGetFieldAccessorTable() { 19899 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable 19900 .ensureFieldAccessorsInitialized( 19901 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.Builder.class); 19902 } 19903 19904 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.newBuilder() 19905 private Builder() { 19906 maybeForceBuilderInitialization(); 19907 } 19908 19909 private Builder( 19910 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19911 super(parent); 19912 maybeForceBuilderInitialization(); 19913 } 19914 private void maybeForceBuilderInitialization() { 19915 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 19916 } 19917 } 19918 private static Builder create() { 19919 return new Builder(); 19920 } 19921 19922 public Builder clear() { 19923 super.clear(); 19924 currentId_ = 0; 19925 bitField0_ = (bitField0_ & ~0x00000001); 19926 tokenSequenceNumber_ = 0; 19927 bitField0_ = (bitField0_ & ~0x00000002); 19928 numKeys_ = 0; 19929 bitField0_ = (bitField0_ & ~0x00000004); 19930 numTokens_ = 0; 19931 bitField0_ = (bitField0_ & ~0x00000008); 19932 return this; 19933 } 19934 19935 public Builder clone() { 19936 return create().mergeFrom(buildPartial()); 19937 } 19938 19939 public com.google.protobuf.Descriptors.Descriptor 19940 getDescriptorForType() { 19941 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 19942 } 19943 19944 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection getDefaultInstanceForType() { 19945 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.getDefaultInstance(); 19946 } 19947 19948 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection build() { 19949 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection result = buildPartial(); 19950 if (!result.isInitialized()) { 19951 throw newUninitializedMessageException(result); 19952 } 19953 return result; 19954 } 19955 19956 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection buildPartial() { 19957 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection(this); 19958 int from_bitField0_ = bitField0_; 19959 int to_bitField0_ = 0; 19960 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 19961 to_bitField0_ |= 0x00000001; 19962 } 19963 result.currentId_ = currentId_; 19964 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 19965 to_bitField0_ |= 0x00000002; 19966 } 19967 result.tokenSequenceNumber_ = tokenSequenceNumber_; 19968 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 19969 to_bitField0_ |= 0x00000004; 19970 } 19971 result.numKeys_ = numKeys_; 19972 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 19973 to_bitField0_ |= 0x00000008; 19974 } 19975 result.numTokens_ = numTokens_; 19976 result.bitField0_ = to_bitField0_; 19977 onBuilt(); 19978 return result; 19979 } 19980 19981 public Builder mergeFrom(com.google.protobuf.Message other) { 19982 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection) { 19983 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection)other); 19984 } else { 19985 super.mergeFrom(other); 19986 return this; 19987 } 19988 } 19989 19990 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection other) { 19991 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.getDefaultInstance()) return this; 19992 if (other.hasCurrentId()) { 19993 setCurrentId(other.getCurrentId()); 19994 } 19995 if (other.hasTokenSequenceNumber()) { 19996 setTokenSequenceNumber(other.getTokenSequenceNumber()); 19997 } 19998 if (other.hasNumKeys()) { 19999 setNumKeys(other.getNumKeys()); 20000 } 20001 if (other.hasNumTokens()) { 20002 setNumTokens(other.getNumTokens()); 20003 } 20004 this.mergeUnknownFields(other.getUnknownFields()); 20005 return this; 20006 } 20007 20008 public final boolean isInitialized() { 20009 return true; 20010 } 20011 20012 public Builder mergeFrom( 20013 com.google.protobuf.CodedInputStream input, 20014 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20015 throws java.io.IOException { 20016 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parsedMessage = null; 20017 try { 20018 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 20019 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 20020 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection) e.getUnfinishedMessage(); 20021 throw e; 20022 } finally { 20023 if (parsedMessage != null) { 20024 mergeFrom(parsedMessage); 20025 } 20026 } 20027 return this; 20028 } 20029 private int bitField0_; 20030 20031 // optional uint32 currentId = 1; 20032 private int currentId_ ; 20033 /** 20034 * <code>optional uint32 currentId = 1;</code> 20035 */ 20036 public boolean hasCurrentId() { 20037 return ((bitField0_ & 0x00000001) == 0x00000001); 20038 } 20039 /** 20040 * <code>optional uint32 currentId = 1;</code> 20041 */ 20042 public int getCurrentId() { 20043 return currentId_; 20044 } 20045 /** 20046 * <code>optional uint32 currentId = 1;</code> 20047 */ 20048 public Builder setCurrentId(int value) { 20049 bitField0_ |= 0x00000001; 20050 currentId_ = value; 20051 onChanged(); 20052 return this; 20053 } 20054 /** 20055 * <code>optional uint32 currentId = 1;</code> 20056 */ 20057 public Builder clearCurrentId() { 20058 bitField0_ = (bitField0_ & ~0x00000001); 20059 currentId_ = 0; 20060 onChanged(); 20061 return this; 20062 } 20063 20064 // optional uint32 tokenSequenceNumber = 2; 20065 private int tokenSequenceNumber_ ; 20066 /** 20067 * <code>optional uint32 tokenSequenceNumber = 2;</code> 20068 */ 20069 public boolean hasTokenSequenceNumber() { 20070 return ((bitField0_ & 0x00000002) == 0x00000002); 20071 } 20072 /** 20073 * <code>optional uint32 tokenSequenceNumber = 2;</code> 20074 */ 20075 public int getTokenSequenceNumber() { 20076 return tokenSequenceNumber_; 20077 } 20078 /** 20079 * <code>optional uint32 tokenSequenceNumber = 2;</code> 20080 */ 20081 public Builder setTokenSequenceNumber(int value) { 20082 bitField0_ |= 0x00000002; 20083 tokenSequenceNumber_ = value; 20084 onChanged(); 20085 return this; 20086 } 20087 /** 20088 * <code>optional uint32 tokenSequenceNumber = 2;</code> 20089 */ 20090 public Builder clearTokenSequenceNumber() { 20091 bitField0_ = (bitField0_ & ~0x00000002); 20092 tokenSequenceNumber_ = 0; 20093 onChanged(); 20094 return this; 20095 } 20096 20097 // optional uint32 numKeys = 3; 20098 private int numKeys_ ; 20099 /** 20100 * <code>optional uint32 numKeys = 3;</code> 20101 */ 20102 public boolean hasNumKeys() { 20103 return ((bitField0_ & 0x00000004) == 0x00000004); 20104 } 20105 /** 20106 * <code>optional uint32 numKeys = 3;</code> 20107 */ 20108 public int getNumKeys() { 20109 return numKeys_; 20110 } 20111 /** 20112 * <code>optional uint32 numKeys = 3;</code> 20113 */ 20114 public Builder setNumKeys(int value) { 20115 bitField0_ |= 0x00000004; 20116 numKeys_ = value; 20117 onChanged(); 20118 return this; 20119 } 20120 /** 20121 * <code>optional uint32 numKeys = 3;</code> 20122 */ 20123 public Builder clearNumKeys() { 20124 bitField0_ = (bitField0_ & ~0x00000004); 20125 numKeys_ = 0; 20126 onChanged(); 20127 return this; 20128 } 20129 20130 // optional uint32 numTokens = 4; 20131 private int numTokens_ ; 20132 /** 20133 * <code>optional uint32 numTokens = 4;</code> 20134 * 20135 * <pre> 20136 * repeated DelegationKey keys 20137 * repeated PersistToken tokens 20138 * </pre> 20139 */ 20140 public boolean hasNumTokens() { 20141 return ((bitField0_ & 0x00000008) == 0x00000008); 20142 } 20143 /** 20144 * <code>optional uint32 numTokens = 4;</code> 20145 * 20146 * <pre> 20147 * repeated DelegationKey keys 20148 * repeated PersistToken tokens 20149 * </pre> 20150 */ 20151 public int getNumTokens() { 20152 return numTokens_; 20153 } 20154 /** 20155 * <code>optional uint32 numTokens = 4;</code> 20156 * 20157 * <pre> 20158 * repeated DelegationKey keys 20159 * repeated PersistToken tokens 20160 * </pre> 20161 */ 20162 public Builder setNumTokens(int value) { 20163 bitField0_ |= 0x00000008; 20164 numTokens_ = value; 20165 onChanged(); 20166 return this; 20167 } 20168 /** 20169 * <code>optional uint32 numTokens = 4;</code> 20170 * 20171 * <pre> 20172 * repeated DelegationKey keys 20173 * repeated PersistToken tokens 20174 * </pre> 20175 */ 20176 public Builder clearNumTokens() { 20177 bitField0_ = (bitField0_ & ~0x00000008); 20178 numTokens_ = 0; 20179 onChanged(); 20180 return this; 20181 } 20182 20183 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection) 20184 } 20185 20186 static { 20187 defaultInstance = new SecretManagerSection(true); 20188 defaultInstance.initFields(); 20189 } 20190 20191 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection) 20192 } 20193 20194 public interface CacheManagerSectionOrBuilder 20195 extends com.google.protobuf.MessageOrBuilder { 20196 20197 // required uint64 nextDirectiveId = 1; 20198 /** 20199 * <code>required uint64 nextDirectiveId = 1;</code> 20200 */ 20201 boolean hasNextDirectiveId(); 20202 /** 20203 * <code>required uint64 nextDirectiveId = 1;</code> 20204 */ 20205 long getNextDirectiveId(); 20206 20207 // required uint32 numPools = 2; 20208 /** 20209 * <code>required uint32 numPools = 2;</code> 20210 */ 20211 boolean hasNumPools(); 20212 /** 20213 * <code>required uint32 numPools = 2;</code> 20214 */ 20215 int getNumPools(); 20216 20217 // required uint32 numDirectives = 3; 20218 /** 20219 * <code>required uint32 numDirectives = 3;</code> 20220 * 20221 * <pre> 20222 * repeated CachePoolInfoProto pools 20223 * repeated CacheDirectiveInfoProto directives 20224 * </pre> 20225 */ 20226 boolean hasNumDirectives(); 20227 /** 20228 * <code>required uint32 numDirectives = 3;</code> 20229 * 20230 * <pre> 20231 * repeated CachePoolInfoProto pools 20232 * repeated CacheDirectiveInfoProto directives 20233 * </pre> 20234 */ 20235 int getNumDirectives(); 20236 } 20237 /** 20238 * Protobuf type {@code hadoop.hdfs.fsimage.CacheManagerSection} 20239 */ 20240 public static final class CacheManagerSection extends 20241 com.google.protobuf.GeneratedMessage 20242 implements CacheManagerSectionOrBuilder { 20243 // Use CacheManagerSection.newBuilder() to construct. 20244 private CacheManagerSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 20245 super(builder); 20246 this.unknownFields = builder.getUnknownFields(); 20247 } 20248 private CacheManagerSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 20249 20250 private static final CacheManagerSection defaultInstance; 20251 public static CacheManagerSection getDefaultInstance() { 20252 return defaultInstance; 20253 } 20254 20255 public CacheManagerSection getDefaultInstanceForType() { 20256 return defaultInstance; 20257 } 20258 20259 private final com.google.protobuf.UnknownFieldSet unknownFields; 20260 @java.lang.Override 20261 public final com.google.protobuf.UnknownFieldSet 20262 getUnknownFields() { 20263 return this.unknownFields; 20264 } 20265 private CacheManagerSection( 20266 com.google.protobuf.CodedInputStream input, 20267 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20268 throws com.google.protobuf.InvalidProtocolBufferException { 20269 initFields(); 20270 int mutable_bitField0_ = 0; 20271 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 20272 com.google.protobuf.UnknownFieldSet.newBuilder(); 20273 try { 20274 boolean done = false; 20275 while (!done) { 20276 int tag = input.readTag(); 20277 switch (tag) { 20278 case 0: 20279 done = true; 20280 break; 20281 default: { 20282 if (!parseUnknownField(input, unknownFields, 20283 extensionRegistry, tag)) { 20284 done = true; 20285 } 20286 break; 20287 } 20288 case 8: { 20289 bitField0_ |= 0x00000001; 20290 nextDirectiveId_ = input.readUInt64(); 20291 break; 20292 } 20293 case 16: { 20294 bitField0_ |= 0x00000002; 20295 numPools_ = input.readUInt32(); 20296 break; 20297 } 20298 case 24: { 20299 bitField0_ |= 0x00000004; 20300 numDirectives_ = input.readUInt32(); 20301 break; 20302 } 20303 } 20304 } 20305 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 20306 throw e.setUnfinishedMessage(this); 20307 } catch (java.io.IOException e) { 20308 throw new com.google.protobuf.InvalidProtocolBufferException( 20309 e.getMessage()).setUnfinishedMessage(this); 20310 } finally { 20311 this.unknownFields = unknownFields.build(); 20312 makeExtensionsImmutable(); 20313 } 20314 } 20315 public static final com.google.protobuf.Descriptors.Descriptor 20316 getDescriptor() { 20317 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 20318 } 20319 20320 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 20321 internalGetFieldAccessorTable() { 20322 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable 20323 .ensureFieldAccessorsInitialized( 20324 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.Builder.class); 20325 } 20326 20327 public static com.google.protobuf.Parser<CacheManagerSection> PARSER = 20328 new com.google.protobuf.AbstractParser<CacheManagerSection>() { 20329 public CacheManagerSection parsePartialFrom( 20330 com.google.protobuf.CodedInputStream input, 20331 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20332 throws com.google.protobuf.InvalidProtocolBufferException { 20333 return new CacheManagerSection(input, extensionRegistry); 20334 } 20335 }; 20336 20337 @java.lang.Override 20338 public com.google.protobuf.Parser<CacheManagerSection> getParserForType() { 20339 return PARSER; 20340 } 20341 20342 private int bitField0_; 20343 // required uint64 nextDirectiveId = 1; 20344 public static final int NEXTDIRECTIVEID_FIELD_NUMBER = 1; 20345 private long nextDirectiveId_; 20346 /** 20347 * <code>required uint64 nextDirectiveId = 1;</code> 20348 */ 20349 public boolean hasNextDirectiveId() { 20350 return ((bitField0_ & 0x00000001) == 0x00000001); 20351 } 20352 /** 20353 * <code>required uint64 nextDirectiveId = 1;</code> 20354 */ 20355 public long getNextDirectiveId() { 20356 return nextDirectiveId_; 20357 } 20358 20359 // required uint32 numPools = 2; 20360 public static final int NUMPOOLS_FIELD_NUMBER = 2; 20361 private int numPools_; 20362 /** 20363 * <code>required uint32 numPools = 2;</code> 20364 */ 20365 public boolean hasNumPools() { 20366 return ((bitField0_ & 0x00000002) == 0x00000002); 20367 } 20368 /** 20369 * <code>required uint32 numPools = 2;</code> 20370 */ 20371 public int getNumPools() { 20372 return numPools_; 20373 } 20374 20375 // required uint32 numDirectives = 3; 20376 public static final int NUMDIRECTIVES_FIELD_NUMBER = 3; 20377 private int numDirectives_; 20378 /** 20379 * <code>required uint32 numDirectives = 3;</code> 20380 * 20381 * <pre> 20382 * repeated CachePoolInfoProto pools 20383 * repeated CacheDirectiveInfoProto directives 20384 * </pre> 20385 */ 20386 public boolean hasNumDirectives() { 20387 return ((bitField0_ & 0x00000004) == 0x00000004); 20388 } 20389 /** 20390 * <code>required uint32 numDirectives = 3;</code> 20391 * 20392 * <pre> 20393 * repeated CachePoolInfoProto pools 20394 * repeated CacheDirectiveInfoProto directives 20395 * </pre> 20396 */ 20397 public int getNumDirectives() { 20398 return numDirectives_; 20399 } 20400 20401 private void initFields() { 20402 nextDirectiveId_ = 0L; 20403 numPools_ = 0; 20404 numDirectives_ = 0; 20405 } 20406 private byte memoizedIsInitialized = -1; 20407 public final boolean isInitialized() { 20408 byte isInitialized = memoizedIsInitialized; 20409 if (isInitialized != -1) return isInitialized == 1; 20410 20411 if (!hasNextDirectiveId()) { 20412 memoizedIsInitialized = 0; 20413 return false; 20414 } 20415 if (!hasNumPools()) { 20416 memoizedIsInitialized = 0; 20417 return false; 20418 } 20419 if (!hasNumDirectives()) { 20420 memoizedIsInitialized = 0; 20421 return false; 20422 } 20423 memoizedIsInitialized = 1; 20424 return true; 20425 } 20426 20427 public void writeTo(com.google.protobuf.CodedOutputStream output) 20428 throws java.io.IOException { 20429 getSerializedSize(); 20430 if (((bitField0_ & 0x00000001) == 0x00000001)) { 20431 output.writeUInt64(1, nextDirectiveId_); 20432 } 20433 if (((bitField0_ & 0x00000002) == 0x00000002)) { 20434 output.writeUInt32(2, numPools_); 20435 } 20436 if (((bitField0_ & 0x00000004) == 0x00000004)) { 20437 output.writeUInt32(3, numDirectives_); 20438 } 20439 getUnknownFields().writeTo(output); 20440 } 20441 20442 private int memoizedSerializedSize = -1; 20443 public int getSerializedSize() { 20444 int size = memoizedSerializedSize; 20445 if (size != -1) return size; 20446 20447 size = 0; 20448 if (((bitField0_ & 0x00000001) == 0x00000001)) { 20449 size += com.google.protobuf.CodedOutputStream 20450 .computeUInt64Size(1, nextDirectiveId_); 20451 } 20452 if (((bitField0_ & 0x00000002) == 0x00000002)) { 20453 size += com.google.protobuf.CodedOutputStream 20454 .computeUInt32Size(2, numPools_); 20455 } 20456 if (((bitField0_ & 0x00000004) == 0x00000004)) { 20457 size += com.google.protobuf.CodedOutputStream 20458 .computeUInt32Size(3, numDirectives_); 20459 } 20460 size += getUnknownFields().getSerializedSize(); 20461 memoizedSerializedSize = size; 20462 return size; 20463 } 20464 20465 private static final long serialVersionUID = 0L; 20466 @java.lang.Override 20467 protected java.lang.Object writeReplace() 20468 throws java.io.ObjectStreamException { 20469 return super.writeReplace(); 20470 } 20471 20472 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20473 com.google.protobuf.ByteString data) 20474 throws com.google.protobuf.InvalidProtocolBufferException { 20475 return PARSER.parseFrom(data); 20476 } 20477 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20478 com.google.protobuf.ByteString data, 20479 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20480 throws com.google.protobuf.InvalidProtocolBufferException { 20481 return PARSER.parseFrom(data, extensionRegistry); 20482 } 20483 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom(byte[] data) 20484 throws com.google.protobuf.InvalidProtocolBufferException { 20485 return PARSER.parseFrom(data); 20486 } 20487 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20488 byte[] data, 20489 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20490 throws com.google.protobuf.InvalidProtocolBufferException { 20491 return PARSER.parseFrom(data, extensionRegistry); 20492 } 20493 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom(java.io.InputStream input) 20494 throws java.io.IOException { 20495 return PARSER.parseFrom(input); 20496 } 20497 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20498 java.io.InputStream input, 20499 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20500 throws java.io.IOException { 20501 return PARSER.parseFrom(input, extensionRegistry); 20502 } 20503 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseDelimitedFrom(java.io.InputStream input) 20504 throws java.io.IOException { 20505 return PARSER.parseDelimitedFrom(input); 20506 } 20507 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseDelimitedFrom( 20508 java.io.InputStream input, 20509 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20510 throws java.io.IOException { 20511 return PARSER.parseDelimitedFrom(input, extensionRegistry); 20512 } 20513 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20514 com.google.protobuf.CodedInputStream input) 20515 throws java.io.IOException { 20516 return PARSER.parseFrom(input); 20517 } 20518 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 20519 com.google.protobuf.CodedInputStream input, 20520 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20521 throws java.io.IOException { 20522 return PARSER.parseFrom(input, extensionRegistry); 20523 } 20524 20525 public static Builder newBuilder() { return Builder.create(); } 20526 public Builder newBuilderForType() { return newBuilder(); } 20527 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection prototype) { 20528 return newBuilder().mergeFrom(prototype); 20529 } 20530 public Builder toBuilder() { return newBuilder(this); } 20531 20532 @java.lang.Override 20533 protected Builder newBuilderForType( 20534 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 20535 Builder builder = new Builder(parent); 20536 return builder; 20537 } 20538 /** 20539 * Protobuf type {@code hadoop.hdfs.fsimage.CacheManagerSection} 20540 */ 20541 public static final class Builder extends 20542 com.google.protobuf.GeneratedMessage.Builder<Builder> 20543 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSectionOrBuilder { 20544 public static final com.google.protobuf.Descriptors.Descriptor 20545 getDescriptor() { 20546 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 20547 } 20548 20549 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 20550 internalGetFieldAccessorTable() { 20551 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable 20552 .ensureFieldAccessorsInitialized( 20553 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.Builder.class); 20554 } 20555 20556 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.newBuilder() 20557 private Builder() { 20558 maybeForceBuilderInitialization(); 20559 } 20560 20561 private Builder( 20562 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 20563 super(parent); 20564 maybeForceBuilderInitialization(); 20565 } 20566 private void maybeForceBuilderInitialization() { 20567 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 20568 } 20569 } 20570 private static Builder create() { 20571 return new Builder(); 20572 } 20573 20574 public Builder clear() { 20575 super.clear(); 20576 nextDirectiveId_ = 0L; 20577 bitField0_ = (bitField0_ & ~0x00000001); 20578 numPools_ = 0; 20579 bitField0_ = (bitField0_ & ~0x00000002); 20580 numDirectives_ = 0; 20581 bitField0_ = (bitField0_ & ~0x00000004); 20582 return this; 20583 } 20584 20585 public Builder clone() { 20586 return create().mergeFrom(buildPartial()); 20587 } 20588 20589 public com.google.protobuf.Descriptors.Descriptor 20590 getDescriptorForType() { 20591 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 20592 } 20593 20594 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection getDefaultInstanceForType() { 20595 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.getDefaultInstance(); 20596 } 20597 20598 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection build() { 20599 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection result = buildPartial(); 20600 if (!result.isInitialized()) { 20601 throw newUninitializedMessageException(result); 20602 } 20603 return result; 20604 } 20605 20606 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection buildPartial() { 20607 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection(this); 20608 int from_bitField0_ = bitField0_; 20609 int to_bitField0_ = 0; 20610 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 20611 to_bitField0_ |= 0x00000001; 20612 } 20613 result.nextDirectiveId_ = nextDirectiveId_; 20614 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 20615 to_bitField0_ |= 0x00000002; 20616 } 20617 result.numPools_ = numPools_; 20618 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 20619 to_bitField0_ |= 0x00000004; 20620 } 20621 result.numDirectives_ = numDirectives_; 20622 result.bitField0_ = to_bitField0_; 20623 onBuilt(); 20624 return result; 20625 } 20626 20627 public Builder mergeFrom(com.google.protobuf.Message other) { 20628 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection) { 20629 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection)other); 20630 } else { 20631 super.mergeFrom(other); 20632 return this; 20633 } 20634 } 20635 20636 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection other) { 20637 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.getDefaultInstance()) return this; 20638 if (other.hasNextDirectiveId()) { 20639 setNextDirectiveId(other.getNextDirectiveId()); 20640 } 20641 if (other.hasNumPools()) { 20642 setNumPools(other.getNumPools()); 20643 } 20644 if (other.hasNumDirectives()) { 20645 setNumDirectives(other.getNumDirectives()); 20646 } 20647 this.mergeUnknownFields(other.getUnknownFields()); 20648 return this; 20649 } 20650 20651 public final boolean isInitialized() { 20652 if (!hasNextDirectiveId()) { 20653 20654 return false; 20655 } 20656 if (!hasNumPools()) { 20657 20658 return false; 20659 } 20660 if (!hasNumDirectives()) { 20661 20662 return false; 20663 } 20664 return true; 20665 } 20666 20667 public Builder mergeFrom( 20668 com.google.protobuf.CodedInputStream input, 20669 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20670 throws java.io.IOException { 20671 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parsedMessage = null; 20672 try { 20673 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 20674 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 20675 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection) e.getUnfinishedMessage(); 20676 throw e; 20677 } finally { 20678 if (parsedMessage != null) { 20679 mergeFrom(parsedMessage); 20680 } 20681 } 20682 return this; 20683 } 20684 private int bitField0_; 20685 20686 // required uint64 nextDirectiveId = 1; 20687 private long nextDirectiveId_ ; 20688 /** 20689 * <code>required uint64 nextDirectiveId = 1;</code> 20690 */ 20691 public boolean hasNextDirectiveId() { 20692 return ((bitField0_ & 0x00000001) == 0x00000001); 20693 } 20694 /** 20695 * <code>required uint64 nextDirectiveId = 1;</code> 20696 */ 20697 public long getNextDirectiveId() { 20698 return nextDirectiveId_; 20699 } 20700 /** 20701 * <code>required uint64 nextDirectiveId = 1;</code> 20702 */ 20703 public Builder setNextDirectiveId(long value) { 20704 bitField0_ |= 0x00000001; 20705 nextDirectiveId_ = value; 20706 onChanged(); 20707 return this; 20708 } 20709 /** 20710 * <code>required uint64 nextDirectiveId = 1;</code> 20711 */ 20712 public Builder clearNextDirectiveId() { 20713 bitField0_ = (bitField0_ & ~0x00000001); 20714 nextDirectiveId_ = 0L; 20715 onChanged(); 20716 return this; 20717 } 20718 20719 // required uint32 numPools = 2; 20720 private int numPools_ ; 20721 /** 20722 * <code>required uint32 numPools = 2;</code> 20723 */ 20724 public boolean hasNumPools() { 20725 return ((bitField0_ & 0x00000002) == 0x00000002); 20726 } 20727 /** 20728 * <code>required uint32 numPools = 2;</code> 20729 */ 20730 public int getNumPools() { 20731 return numPools_; 20732 } 20733 /** 20734 * <code>required uint32 numPools = 2;</code> 20735 */ 20736 public Builder setNumPools(int value) { 20737 bitField0_ |= 0x00000002; 20738 numPools_ = value; 20739 onChanged(); 20740 return this; 20741 } 20742 /** 20743 * <code>required uint32 numPools = 2;</code> 20744 */ 20745 public Builder clearNumPools() { 20746 bitField0_ = (bitField0_ & ~0x00000002); 20747 numPools_ = 0; 20748 onChanged(); 20749 return this; 20750 } 20751 20752 // required uint32 numDirectives = 3; 20753 private int numDirectives_ ; 20754 /** 20755 * <code>required uint32 numDirectives = 3;</code> 20756 * 20757 * <pre> 20758 * repeated CachePoolInfoProto pools 20759 * repeated CacheDirectiveInfoProto directives 20760 * </pre> 20761 */ 20762 public boolean hasNumDirectives() { 20763 return ((bitField0_ & 0x00000004) == 0x00000004); 20764 } 20765 /** 20766 * <code>required uint32 numDirectives = 3;</code> 20767 * 20768 * <pre> 20769 * repeated CachePoolInfoProto pools 20770 * repeated CacheDirectiveInfoProto directives 20771 * </pre> 20772 */ 20773 public int getNumDirectives() { 20774 return numDirectives_; 20775 } 20776 /** 20777 * <code>required uint32 numDirectives = 3;</code> 20778 * 20779 * <pre> 20780 * repeated CachePoolInfoProto pools 20781 * repeated CacheDirectiveInfoProto directives 20782 * </pre> 20783 */ 20784 public Builder setNumDirectives(int value) { 20785 bitField0_ |= 0x00000004; 20786 numDirectives_ = value; 20787 onChanged(); 20788 return this; 20789 } 20790 /** 20791 * <code>required uint32 numDirectives = 3;</code> 20792 * 20793 * <pre> 20794 * repeated CachePoolInfoProto pools 20795 * repeated CacheDirectiveInfoProto directives 20796 * </pre> 20797 */ 20798 public Builder clearNumDirectives() { 20799 bitField0_ = (bitField0_ & ~0x00000004); 20800 numDirectives_ = 0; 20801 onChanged(); 20802 return this; 20803 } 20804 20805 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.CacheManagerSection) 20806 } 20807 20808 static { 20809 defaultInstance = new CacheManagerSection(true); 20810 defaultInstance.initFields(); 20811 } 20812 20813 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.CacheManagerSection) 20814 } 20815 20816 private static com.google.protobuf.Descriptors.Descriptor 20817 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 20818 private static 20819 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20820 internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable; 20821 private static com.google.protobuf.Descriptors.Descriptor 20822 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 20823 private static 20824 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20825 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable; 20826 private static com.google.protobuf.Descriptors.Descriptor 20827 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 20828 private static 20829 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20830 internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable; 20831 private static com.google.protobuf.Descriptors.Descriptor 20832 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 20833 private static 20834 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20835 internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable; 20836 private static com.google.protobuf.Descriptors.Descriptor 20837 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 20838 private static 20839 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20840 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable; 20841 private static com.google.protobuf.Descriptors.Descriptor 20842 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 20843 private static 20844 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20845 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable; 20846 private static com.google.protobuf.Descriptors.Descriptor 20847 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 20848 private static 20849 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20850 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable; 20851 private static com.google.protobuf.Descriptors.Descriptor 20852 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 20853 private static 20854 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20855 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable; 20856 private static com.google.protobuf.Descriptors.Descriptor 20857 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 20858 private static 20859 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20860 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable; 20861 private static com.google.protobuf.Descriptors.Descriptor 20862 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 20863 private static 20864 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20865 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable; 20866 private static com.google.protobuf.Descriptors.Descriptor 20867 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 20868 private static 20869 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20870 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable; 20871 private static com.google.protobuf.Descriptors.Descriptor 20872 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 20873 private static 20874 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20875 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable; 20876 private static com.google.protobuf.Descriptors.Descriptor 20877 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 20878 private static 20879 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20880 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable; 20881 private static com.google.protobuf.Descriptors.Descriptor 20882 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 20883 private static 20884 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20885 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable; 20886 private static com.google.protobuf.Descriptors.Descriptor 20887 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 20888 private static 20889 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20890 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable; 20891 private static com.google.protobuf.Descriptors.Descriptor 20892 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 20893 private static 20894 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20895 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable; 20896 private static com.google.protobuf.Descriptors.Descriptor 20897 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 20898 private static 20899 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20900 internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable; 20901 private static com.google.protobuf.Descriptors.Descriptor 20902 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 20903 private static 20904 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20905 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable; 20906 private static com.google.protobuf.Descriptors.Descriptor 20907 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 20908 private static 20909 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20910 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable; 20911 private static com.google.protobuf.Descriptors.Descriptor 20912 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 20913 private static 20914 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20915 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable; 20916 private static com.google.protobuf.Descriptors.Descriptor 20917 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 20918 private static 20919 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20920 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable; 20921 private static com.google.protobuf.Descriptors.Descriptor 20922 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 20923 private static 20924 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20925 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable; 20926 private static com.google.protobuf.Descriptors.Descriptor 20927 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 20928 private static 20929 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20930 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable; 20931 private static com.google.protobuf.Descriptors.Descriptor 20932 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 20933 private static 20934 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20935 internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable; 20936 private static com.google.protobuf.Descriptors.Descriptor 20937 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 20938 private static 20939 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20940 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable; 20941 private static com.google.protobuf.Descriptors.Descriptor 20942 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 20943 private static 20944 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20945 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable; 20946 private static com.google.protobuf.Descriptors.Descriptor 20947 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 20948 private static 20949 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20950 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable; 20951 private static com.google.protobuf.Descriptors.Descriptor 20952 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 20953 private static 20954 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20955 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable; 20956 private static com.google.protobuf.Descriptors.Descriptor 20957 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 20958 private static 20959 com.google.protobuf.GeneratedMessage.FieldAccessorTable 20960 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable; 20961 20962 public static com.google.protobuf.Descriptors.FileDescriptor 20963 getDescriptor() { 20964 return descriptor; 20965 } 20966 private static com.google.protobuf.Descriptors.FileDescriptor 20967 descriptor; 20968 static { 20969 java.lang.String[] descriptorData = { 20970 "\n\rfsimage.proto\022\023hadoop.hdfs.fsimage\032\nhd" + 20971 "fs.proto\032\tacl.proto\"\277\001\n\013FileSummary\022\025\n\ro" + 20972 "ndiskVersion\030\001 \002(\r\022\025\n\rlayoutVersion\030\002 \002(" + 20973 "\r\022\r\n\005codec\030\003 \001(\t\022:\n\010sections\030\004 \003(\0132(.had" + 20974 "oop.hdfs.fsimage.FileSummary.Section\0327\n\007" + 20975 "Section\022\014\n\004name\030\001 \001(\t\022\016\n\006length\030\002 \001(\004\022\016\n" + 20976 "\006offset\030\003 \001(\004\"\277\001\n\021NameSystemSection\022\023\n\013n" + 20977 "amespaceId\030\001 \001(\r\022\022\n\ngenstampV1\030\002 \001(\004\022\022\n\n" + 20978 "genstampV2\030\003 \001(\004\022\027\n\017genstampV1Limit\030\004 \001(" + 20979 "\004\022\034\n\024lastAllocatedBlockId\030\005 \001(\004\022\025\n\rtrans", 20980 "actionId\030\006 \001(\004\022\037\n\027rollingUpgradeStartTim" + 20981 "e\030\007 \001(\004\"\267\010\n\014INodeSection\022\023\n\013lastInodeId\030" + 20982 "\001 \001(\004\022\021\n\tnumInodes\030\002 \001(\004\032I\n\034FileUnderCon" + 20983 "structionFeature\022\022\n\nclientName\030\001 \001(\t\022\025\n\r" + 20984 "clientMachine\030\002 \001(\t\032&\n\017AclFeatureProto\022\023" + 20985 "\n\007entries\030\002 \003(\007B\002\020\001\032\267\002\n\tINodeFile\022\023\n\013rep" + 20986 "lication\030\001 \001(\r\022\030\n\020modificationTime\030\002 \001(\004" + 20987 "\022\022\n\naccessTime\030\003 \001(\004\022\032\n\022preferredBlockSi" + 20988 "ze\030\004 \001(\004\022\022\n\npermission\030\005 \001(\006\022\'\n\006blocks\030\006" + 20989 " \003(\0132\027.hadoop.hdfs.BlockProto\022N\n\006fileUC\030", 20990 "\007 \001(\0132>.hadoop.hdfs.fsimage.INodeSection" + 20991 ".FileUnderConstructionFeature\022>\n\003acl\030\010 \001" + 20992 "(\01321.hadoop.hdfs.fsimage.INodeSection.Ac" + 20993 "lFeatureProto\032\240\001\n\016INodeDirectory\022\030\n\020modi" + 20994 "ficationTime\030\001 \001(\004\022\017\n\007nsQuota\030\002 \001(\004\022\017\n\007d" + 20995 "sQuota\030\003 \001(\004\022\022\n\npermission\030\004 \001(\006\022>\n\003acl\030" + 20996 "\005 \001(\01321.hadoop.hdfs.fsimage.INodeSection" + 20997 ".AclFeatureProto\032`\n\014INodeSymlink\022\022\n\nperm" + 20998 "ission\030\001 \001(\006\022\016\n\006target\030\002 \001(\014\022\030\n\020modifica" + 20999 "tionTime\030\003 \001(\004\022\022\n\naccessTime\030\004 \001(\004\032\314\002\n\005I", 21000 "Node\022:\n\004type\030\001 \002(\0162,.hadoop.hdfs.fsimage" + 21001 ".INodeSection.INode.Type\022\n\n\002id\030\002 \002(\004\022\014\n\004" + 21002 "name\030\003 \001(\014\0229\n\004file\030\004 \001(\0132+.hadoop.hdfs.f" + 21003 "simage.INodeSection.INodeFile\022C\n\tdirecto" + 21004 "ry\030\005 \001(\01320.hadoop.hdfs.fsimage.INodeSect" + 21005 "ion.INodeDirectory\022?\n\007symlink\030\006 \001(\0132..ha" + 21006 "doop.hdfs.fsimage.INodeSection.INodeSyml" + 21007 "ink\",\n\004Type\022\010\n\004FILE\020\001\022\r\n\tDIRECTORY\020\002\022\013\n\007" + 21008 "SYMLINK\020\003\"`\n\035FilesUnderConstructionSecti" + 21009 "on\032?\n\032FileUnderConstructionEntry\022\017\n\007inod", 21010 "eId\030\001 \001(\004\022\020\n\010fullPath\030\002 \001(\t\"b\n\025INodeDire" + 21011 "ctorySection\032I\n\010DirEntry\022\016\n\006parent\030\001 \001(\004" + 21012 "\022\024\n\010children\030\002 \003(\004B\002\020\001\022\027\n\013refChildren\030\003 " + 21013 "\003(\rB\002\020\001\"z\n\025INodeReferenceSection\032a\n\016INod" + 21014 "eReference\022\022\n\nreferredId\030\001 \001(\004\022\014\n\004name\030\002" + 21015 " \001(\014\022\025\n\rdstSnapshotId\030\003 \001(\r\022\026\n\016lastSnaps" + 21016 "hotId\030\004 \001(\r\"\265\001\n\017SnapshotSection\022\027\n\017snaps" + 21017 "hotCounter\030\001 \001(\r\022\034\n\020snapshottableDir\030\002 \003" + 21018 "(\004B\002\020\001\022\024\n\014numSnapshots\030\003 \001(\r\032U\n\010Snapshot" + 21019 "\022\022\n\nsnapshotId\030\001 \001(\r\0225\n\004root\030\002 \001(\0132\'.had", 21020 "oop.hdfs.fsimage.INodeSection.INode\"\327\004\n\023" + 21021 "SnapshotDiffSection\032 \n\020CreatedListEntry\022" + 21022 "\014\n\004name\030\001 \001(\014\032\367\001\n\rDirectoryDiff\022\022\n\nsnaps" + 21023 "hotId\030\001 \001(\r\022\024\n\014childrenSize\030\002 \001(\r\022\026\n\016isS" + 21024 "napshotRoot\030\003 \001(\010\022\014\n\004name\030\004 \001(\014\022F\n\014snaps" + 21025 "hotCopy\030\005 \001(\01320.hadoop.hdfs.fsimage.INod" + 21026 "eSection.INodeDirectory\022\027\n\017createdListSi" + 21027 "ze\030\006 \001(\r\022\030\n\014deletedINode\030\007 \003(\004B\002\020\001\022\033\n\017de" + 21028 "letedINodeRef\030\010 \003(\rB\002\020\001\032\201\001\n\010FileDiff\022\022\n\n" + 21029 "snapshotId\030\001 \001(\r\022\020\n\010fileSize\030\002 \001(\004\022\014\n\004na", 21030 "me\030\003 \001(\014\022A\n\014snapshotCopy\030\004 \001(\0132+.hadoop." + 21031 "hdfs.fsimage.INodeSection.INodeFile\032\237\001\n\t" + 21032 "DiffEntry\022E\n\004type\030\001 \002(\01627.hadoop.hdfs.fs" + 21033 "image.SnapshotDiffSection.DiffEntry.Type" + 21034 "\022\017\n\007inodeId\030\002 \001(\004\022\021\n\tnumOfDiff\030\003 \001(\r\"\'\n\004" + 21035 "Type\022\014\n\010FILEDIFF\020\001\022\021\n\rDIRECTORYDIFF\020\002\"H\n" + 21036 "\022StringTableSection\022\020\n\010numEntry\030\001 \001(\r\032 \n" + 21037 "\005Entry\022\n\n\002id\030\001 \001(\r\022\013\n\003str\030\002 \001(\t\"\341\002\n\024Secr" + 21038 "etManagerSection\022\021\n\tcurrentId\030\001 \001(\r\022\033\n\023t" + 21039 "okenSequenceNumber\030\002 \001(\r\022\017\n\007numKeys\030\003 \001(", 21040 "\r\022\021\n\tnumTokens\030\004 \001(\r\032<\n\rDelegationKey\022\n\n" + 21041 "\002id\030\001 \001(\r\022\022\n\nexpiryDate\030\002 \001(\004\022\013\n\003key\030\003 \001" + 21042 "(\014\032\266\001\n\014PersistToken\022\017\n\007version\030\001 \001(\r\022\r\n\005" + 21043 "owner\030\002 \001(\t\022\017\n\007renewer\030\003 \001(\t\022\020\n\010realUser" + 21044 "\030\004 \001(\t\022\021\n\tissueDate\030\005 \001(\004\022\017\n\007maxDate\030\006 \001" + 21045 "(\004\022\026\n\016sequenceNumber\030\007 \001(\r\022\023\n\013masterKeyI" + 21046 "d\030\010 \001(\r\022\022\n\nexpiryDate\030\t \001(\004\"W\n\023CacheMana" + 21047 "gerSection\022\027\n\017nextDirectiveId\030\001 \002(\004\022\020\n\010n" + 21048 "umPools\030\002 \002(\r\022\025\n\rnumDirectives\030\003 \002(\rB6\n&" + 21049 "org.apache.hadoop.hdfs.server.namenodeB\014", 21050 "FsImageProto" 21051 }; 21052 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 21053 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 21054 public com.google.protobuf.ExtensionRegistry assignDescriptors( 21055 com.google.protobuf.Descriptors.FileDescriptor root) { 21056 descriptor = root; 21057 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor = 21058 getDescriptor().getMessageTypes().get(0); 21059 internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable = new 21060 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21061 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor, 21062 new java.lang.String[] { "OndiskVersion", "LayoutVersion", "Codec", "Sections", }); 21063 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor = 21064 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor.getNestedTypes().get(0); 21065 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable = new 21066 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21067 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor, 21068 new java.lang.String[] { "Name", "Length", "Offset", }); 21069 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor = 21070 getDescriptor().getMessageTypes().get(1); 21071 internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable = new 21072 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21073 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor, 21074 new java.lang.String[] { "NamespaceId", "GenstampV1", "GenstampV2", "GenstampV1Limit", "LastAllocatedBlockId", "TransactionId", "RollingUpgradeStartTime", }); 21075 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor = 21076 getDescriptor().getMessageTypes().get(2); 21077 internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable = new 21078 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21079 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor, 21080 new java.lang.String[] { "LastInodeId", "NumInodes", }); 21081 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor = 21082 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(0); 21083 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable = new 21084 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21085 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor, 21086 new java.lang.String[] { "ClientName", "ClientMachine", }); 21087 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor = 21088 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(1); 21089 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable = new 21090 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21091 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor, 21092 new java.lang.String[] { "Entries", }); 21093 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor = 21094 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(2); 21095 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable = new 21096 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21097 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor, 21098 new java.lang.String[] { "Replication", "ModificationTime", "AccessTime", "PreferredBlockSize", "Permission", "Blocks", "FileUC", "Acl", }); 21099 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor = 21100 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(3); 21101 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable = new 21102 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21103 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor, 21104 new java.lang.String[] { "ModificationTime", "NsQuota", "DsQuota", "Permission", "Acl", }); 21105 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor = 21106 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(4); 21107 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable = new 21108 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21109 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor, 21110 new java.lang.String[] { "Permission", "Target", "ModificationTime", "AccessTime", }); 21111 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor = 21112 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(5); 21113 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable = new 21114 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21115 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor, 21116 new java.lang.String[] { "Type", "Id", "Name", "File", "Directory", "Symlink", }); 21117 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor = 21118 getDescriptor().getMessageTypes().get(3); 21119 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable = new 21120 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21121 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor, 21122 new java.lang.String[] { }); 21123 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor = 21124 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor.getNestedTypes().get(0); 21125 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable = new 21126 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21127 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor, 21128 new java.lang.String[] { "InodeId", "FullPath", }); 21129 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor = 21130 getDescriptor().getMessageTypes().get(4); 21131 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable = new 21132 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21133 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor, 21134 new java.lang.String[] { }); 21135 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor = 21136 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor.getNestedTypes().get(0); 21137 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable = new 21138 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21139 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor, 21140 new java.lang.String[] { "Parent", "Children", "RefChildren", }); 21141 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor = 21142 getDescriptor().getMessageTypes().get(5); 21143 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable = new 21144 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21145 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor, 21146 new java.lang.String[] { }); 21147 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor = 21148 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor.getNestedTypes().get(0); 21149 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable = new 21150 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21151 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor, 21152 new java.lang.String[] { "ReferredId", "Name", "DstSnapshotId", "LastSnapshotId", }); 21153 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor = 21154 getDescriptor().getMessageTypes().get(6); 21155 internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable = new 21156 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21157 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor, 21158 new java.lang.String[] { "SnapshotCounter", "SnapshottableDir", "NumSnapshots", }); 21159 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor = 21160 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor.getNestedTypes().get(0); 21161 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable = new 21162 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21163 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor, 21164 new java.lang.String[] { "SnapshotId", "Root", }); 21165 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor = 21166 getDescriptor().getMessageTypes().get(7); 21167 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable = new 21168 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21169 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor, 21170 new java.lang.String[] { }); 21171 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor = 21172 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(0); 21173 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable = new 21174 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21175 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor, 21176 new java.lang.String[] { "Name", }); 21177 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor = 21178 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(1); 21179 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable = new 21180 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21181 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor, 21182 new java.lang.String[] { "SnapshotId", "ChildrenSize", "IsSnapshotRoot", "Name", "SnapshotCopy", "CreatedListSize", "DeletedINode", "DeletedINodeRef", }); 21183 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor = 21184 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(2); 21185 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable = new 21186 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21187 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor, 21188 new java.lang.String[] { "SnapshotId", "FileSize", "Name", "SnapshotCopy", }); 21189 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor = 21190 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(3); 21191 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable = new 21192 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21193 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor, 21194 new java.lang.String[] { "Type", "InodeId", "NumOfDiff", }); 21195 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor = 21196 getDescriptor().getMessageTypes().get(8); 21197 internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable = new 21198 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21199 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor, 21200 new java.lang.String[] { "NumEntry", }); 21201 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor = 21202 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor.getNestedTypes().get(0); 21203 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable = new 21204 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21205 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor, 21206 new java.lang.String[] { "Id", "Str", }); 21207 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor = 21208 getDescriptor().getMessageTypes().get(9); 21209 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable = new 21210 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21211 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor, 21212 new java.lang.String[] { "CurrentId", "TokenSequenceNumber", "NumKeys", "NumTokens", }); 21213 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor = 21214 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor.getNestedTypes().get(0); 21215 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable = new 21216 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21217 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor, 21218 new java.lang.String[] { "Id", "ExpiryDate", "Key", }); 21219 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor = 21220 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor.getNestedTypes().get(1); 21221 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable = new 21222 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21223 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor, 21224 new java.lang.String[] { "Version", "Owner", "Renewer", "RealUser", "IssueDate", "MaxDate", "SequenceNumber", "MasterKeyId", "ExpiryDate", }); 21225 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor = 21226 getDescriptor().getMessageTypes().get(10); 21227 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable = new 21228 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 21229 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor, 21230 new java.lang.String[] { "NextDirectiveId", "NumPools", "NumDirectives", }); 21231 return null; 21232 } 21233 }; 21234 com.google.protobuf.Descriptors.FileDescriptor 21235 .internalBuildGeneratedFileFrom(descriptorData, 21236 new com.google.protobuf.Descriptors.FileDescriptor[] { 21237 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(), 21238 org.apache.hadoop.hdfs.protocol.proto.AclProtos.getDescriptor(), 21239 }, assigner); 21240 } 21241 21242 // @@protoc_insertion_point(outer_class_scope) 21243 }