001 // Generated by the protocol buffer compiler. DO NOT EDIT! 002 // source: fsimage.proto 003 004 package org.apache.hadoop.hdfs.server.namenode; 005 006 public final class FsImageProto { 007 private FsImageProto() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public interface FileSummaryOrBuilder 012 extends com.google.protobuf.MessageOrBuilder { 013 014 // required uint32 ondiskVersion = 1; 015 /** 016 * <code>required uint32 ondiskVersion = 1;</code> 017 * 018 * <pre> 019 * The version of the above EBNF grammars. 020 * </pre> 021 */ 022 boolean hasOndiskVersion(); 023 /** 024 * <code>required uint32 ondiskVersion = 1;</code> 025 * 026 * <pre> 027 * The version of the above EBNF grammars. 028 * </pre> 029 */ 030 int getOndiskVersion(); 031 032 // required uint32 layoutVersion = 2; 033 /** 034 * <code>required uint32 layoutVersion = 2;</code> 035 * 036 * <pre> 037 * layoutVersion describes which features are available in the 038 * FSImage. 039 * </pre> 040 */ 041 boolean hasLayoutVersion(); 042 /** 043 * <code>required uint32 layoutVersion = 2;</code> 044 * 045 * <pre> 046 * layoutVersion describes which features are available in the 047 * FSImage. 048 * </pre> 049 */ 050 int getLayoutVersion(); 051 052 // optional string codec = 3; 053 /** 054 * <code>optional string codec = 3;</code> 055 */ 056 boolean hasCodec(); 057 /** 058 * <code>optional string codec = 3;</code> 059 */ 060 java.lang.String getCodec(); 061 /** 062 * <code>optional string codec = 3;</code> 063 */ 064 com.google.protobuf.ByteString 065 getCodecBytes(); 066 067 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 068 /** 069 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 070 */ 071 java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> 072 getSectionsList(); 073 /** 074 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 075 */ 076 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index); 077 /** 078 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 079 */ 080 int getSectionsCount(); 081 /** 082 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 083 */ 084 java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 085 getSectionsOrBuilderList(); 086 /** 087 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 088 */ 089 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 090 int index); 091 } 092 /** 093 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary} 094 */ 095 public static final class FileSummary extends 096 com.google.protobuf.GeneratedMessage 097 implements FileSummaryOrBuilder { 098 // Use FileSummary.newBuilder() to construct. 099 private FileSummary(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 100 super(builder); 101 this.unknownFields = builder.getUnknownFields(); 102 } 103 private FileSummary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 104 105 private static final FileSummary defaultInstance; 106 public static FileSummary getDefaultInstance() { 107 return defaultInstance; 108 } 109 110 public FileSummary getDefaultInstanceForType() { 111 return defaultInstance; 112 } 113 114 private final com.google.protobuf.UnknownFieldSet unknownFields; 115 @java.lang.Override 116 public final com.google.protobuf.UnknownFieldSet 117 getUnknownFields() { 118 return this.unknownFields; 119 } 120 private FileSummary( 121 com.google.protobuf.CodedInputStream input, 122 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 123 throws com.google.protobuf.InvalidProtocolBufferException { 124 initFields(); 125 int mutable_bitField0_ = 0; 126 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 127 com.google.protobuf.UnknownFieldSet.newBuilder(); 128 try { 129 boolean done = false; 130 while (!done) { 131 int tag = input.readTag(); 132 switch (tag) { 133 case 0: 134 done = true; 135 break; 136 default: { 137 if (!parseUnknownField(input, unknownFields, 138 extensionRegistry, tag)) { 139 done = true; 140 } 141 break; 142 } 143 case 8: { 144 bitField0_ |= 0x00000001; 145 ondiskVersion_ = input.readUInt32(); 146 break; 147 } 148 case 16: { 149 bitField0_ |= 0x00000002; 150 layoutVersion_ = input.readUInt32(); 151 break; 152 } 153 case 26: { 154 bitField0_ |= 0x00000004; 155 codec_ = input.readBytes(); 156 break; 157 } 158 case 34: { 159 if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) { 160 sections_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section>(); 161 mutable_bitField0_ |= 0x00000008; 162 } 163 sections_.add(input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.PARSER, extensionRegistry)); 164 break; 165 } 166 } 167 } 168 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 169 throw e.setUnfinishedMessage(this); 170 } catch (java.io.IOException e) { 171 throw new com.google.protobuf.InvalidProtocolBufferException( 172 e.getMessage()).setUnfinishedMessage(this); 173 } finally { 174 if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) { 175 sections_ = java.util.Collections.unmodifiableList(sections_); 176 } 177 this.unknownFields = unknownFields.build(); 178 makeExtensionsImmutable(); 179 } 180 } 181 public static final com.google.protobuf.Descriptors.Descriptor 182 getDescriptor() { 183 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 184 } 185 186 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 187 internalGetFieldAccessorTable() { 188 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable 189 .ensureFieldAccessorsInitialized( 190 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Builder.class); 191 } 192 193 public static com.google.protobuf.Parser<FileSummary> PARSER = 194 new com.google.protobuf.AbstractParser<FileSummary>() { 195 public FileSummary parsePartialFrom( 196 com.google.protobuf.CodedInputStream input, 197 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 198 throws com.google.protobuf.InvalidProtocolBufferException { 199 return new FileSummary(input, extensionRegistry); 200 } 201 }; 202 203 @java.lang.Override 204 public com.google.protobuf.Parser<FileSummary> getParserForType() { 205 return PARSER; 206 } 207 208 public interface SectionOrBuilder 209 extends com.google.protobuf.MessageOrBuilder { 210 211 // optional string name = 1; 212 /** 213 * <code>optional string name = 1;</code> 214 */ 215 boolean hasName(); 216 /** 217 * <code>optional string name = 1;</code> 218 */ 219 java.lang.String getName(); 220 /** 221 * <code>optional string name = 1;</code> 222 */ 223 com.google.protobuf.ByteString 224 getNameBytes(); 225 226 // optional uint64 length = 2; 227 /** 228 * <code>optional uint64 length = 2;</code> 229 */ 230 boolean hasLength(); 231 /** 232 * <code>optional uint64 length = 2;</code> 233 */ 234 long getLength(); 235 236 // optional uint64 offset = 3; 237 /** 238 * <code>optional uint64 offset = 3;</code> 239 */ 240 boolean hasOffset(); 241 /** 242 * <code>optional uint64 offset = 3;</code> 243 */ 244 long getOffset(); 245 } 246 /** 247 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary.Section} 248 * 249 * <pre> 250 * index for each section 251 * </pre> 252 */ 253 public static final class Section extends 254 com.google.protobuf.GeneratedMessage 255 implements SectionOrBuilder { 256 // Use Section.newBuilder() to construct. 257 private Section(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 258 super(builder); 259 this.unknownFields = builder.getUnknownFields(); 260 } 261 private Section(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 262 263 private static final Section defaultInstance; 264 public static Section getDefaultInstance() { 265 return defaultInstance; 266 } 267 268 public Section getDefaultInstanceForType() { 269 return defaultInstance; 270 } 271 272 private final com.google.protobuf.UnknownFieldSet unknownFields; 273 @java.lang.Override 274 public final com.google.protobuf.UnknownFieldSet 275 getUnknownFields() { 276 return this.unknownFields; 277 } 278 private Section( 279 com.google.protobuf.CodedInputStream input, 280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 281 throws com.google.protobuf.InvalidProtocolBufferException { 282 initFields(); 283 int mutable_bitField0_ = 0; 284 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 285 com.google.protobuf.UnknownFieldSet.newBuilder(); 286 try { 287 boolean done = false; 288 while (!done) { 289 int tag = input.readTag(); 290 switch (tag) { 291 case 0: 292 done = true; 293 break; 294 default: { 295 if (!parseUnknownField(input, unknownFields, 296 extensionRegistry, tag)) { 297 done = true; 298 } 299 break; 300 } 301 case 10: { 302 bitField0_ |= 0x00000001; 303 name_ = input.readBytes(); 304 break; 305 } 306 case 16: { 307 bitField0_ |= 0x00000002; 308 length_ = input.readUInt64(); 309 break; 310 } 311 case 24: { 312 bitField0_ |= 0x00000004; 313 offset_ = input.readUInt64(); 314 break; 315 } 316 } 317 } 318 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 319 throw e.setUnfinishedMessage(this); 320 } catch (java.io.IOException e) { 321 throw new com.google.protobuf.InvalidProtocolBufferException( 322 e.getMessage()).setUnfinishedMessage(this); 323 } finally { 324 this.unknownFields = unknownFields.build(); 325 makeExtensionsImmutable(); 326 } 327 } 328 public static final com.google.protobuf.Descriptors.Descriptor 329 getDescriptor() { 330 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 331 } 332 333 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 334 internalGetFieldAccessorTable() { 335 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable 336 .ensureFieldAccessorsInitialized( 337 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder.class); 338 } 339 340 public static com.google.protobuf.Parser<Section> PARSER = 341 new com.google.protobuf.AbstractParser<Section>() { 342 public Section parsePartialFrom( 343 com.google.protobuf.CodedInputStream input, 344 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 345 throws com.google.protobuf.InvalidProtocolBufferException { 346 return new Section(input, extensionRegistry); 347 } 348 }; 349 350 @java.lang.Override 351 public com.google.protobuf.Parser<Section> getParserForType() { 352 return PARSER; 353 } 354 355 private int bitField0_; 356 // optional string name = 1; 357 public static final int NAME_FIELD_NUMBER = 1; 358 private java.lang.Object name_; 359 /** 360 * <code>optional string name = 1;</code> 361 */ 362 public boolean hasName() { 363 return ((bitField0_ & 0x00000001) == 0x00000001); 364 } 365 /** 366 * <code>optional string name = 1;</code> 367 */ 368 public java.lang.String getName() { 369 java.lang.Object ref = name_; 370 if (ref instanceof java.lang.String) { 371 return (java.lang.String) ref; 372 } else { 373 com.google.protobuf.ByteString bs = 374 (com.google.protobuf.ByteString) ref; 375 java.lang.String s = bs.toStringUtf8(); 376 if (bs.isValidUtf8()) { 377 name_ = s; 378 } 379 return s; 380 } 381 } 382 /** 383 * <code>optional string name = 1;</code> 384 */ 385 public com.google.protobuf.ByteString 386 getNameBytes() { 387 java.lang.Object ref = name_; 388 if (ref instanceof java.lang.String) { 389 com.google.protobuf.ByteString b = 390 com.google.protobuf.ByteString.copyFromUtf8( 391 (java.lang.String) ref); 392 name_ = b; 393 return b; 394 } else { 395 return (com.google.protobuf.ByteString) ref; 396 } 397 } 398 399 // optional uint64 length = 2; 400 public static final int LENGTH_FIELD_NUMBER = 2; 401 private long length_; 402 /** 403 * <code>optional uint64 length = 2;</code> 404 */ 405 public boolean hasLength() { 406 return ((bitField0_ & 0x00000002) == 0x00000002); 407 } 408 /** 409 * <code>optional uint64 length = 2;</code> 410 */ 411 public long getLength() { 412 return length_; 413 } 414 415 // optional uint64 offset = 3; 416 public static final int OFFSET_FIELD_NUMBER = 3; 417 private long offset_; 418 /** 419 * <code>optional uint64 offset = 3;</code> 420 */ 421 public boolean hasOffset() { 422 return ((bitField0_ & 0x00000004) == 0x00000004); 423 } 424 /** 425 * <code>optional uint64 offset = 3;</code> 426 */ 427 public long getOffset() { 428 return offset_; 429 } 430 431 private void initFields() { 432 name_ = ""; 433 length_ = 0L; 434 offset_ = 0L; 435 } 436 private byte memoizedIsInitialized = -1; 437 public final boolean isInitialized() { 438 byte isInitialized = memoizedIsInitialized; 439 if (isInitialized != -1) return isInitialized == 1; 440 441 memoizedIsInitialized = 1; 442 return true; 443 } 444 445 public void writeTo(com.google.protobuf.CodedOutputStream output) 446 throws java.io.IOException { 447 getSerializedSize(); 448 if (((bitField0_ & 0x00000001) == 0x00000001)) { 449 output.writeBytes(1, getNameBytes()); 450 } 451 if (((bitField0_ & 0x00000002) == 0x00000002)) { 452 output.writeUInt64(2, length_); 453 } 454 if (((bitField0_ & 0x00000004) == 0x00000004)) { 455 output.writeUInt64(3, offset_); 456 } 457 getUnknownFields().writeTo(output); 458 } 459 460 private int memoizedSerializedSize = -1; 461 public int getSerializedSize() { 462 int size = memoizedSerializedSize; 463 if (size != -1) return size; 464 465 size = 0; 466 if (((bitField0_ & 0x00000001) == 0x00000001)) { 467 size += com.google.protobuf.CodedOutputStream 468 .computeBytesSize(1, getNameBytes()); 469 } 470 if (((bitField0_ & 0x00000002) == 0x00000002)) { 471 size += com.google.protobuf.CodedOutputStream 472 .computeUInt64Size(2, length_); 473 } 474 if (((bitField0_ & 0x00000004) == 0x00000004)) { 475 size += com.google.protobuf.CodedOutputStream 476 .computeUInt64Size(3, offset_); 477 } 478 size += getUnknownFields().getSerializedSize(); 479 memoizedSerializedSize = size; 480 return size; 481 } 482 483 private static final long serialVersionUID = 0L; 484 @java.lang.Override 485 protected java.lang.Object writeReplace() 486 throws java.io.ObjectStreamException { 487 return super.writeReplace(); 488 } 489 490 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 491 com.google.protobuf.ByteString data) 492 throws com.google.protobuf.InvalidProtocolBufferException { 493 return PARSER.parseFrom(data); 494 } 495 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 496 com.google.protobuf.ByteString data, 497 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 498 throws com.google.protobuf.InvalidProtocolBufferException { 499 return PARSER.parseFrom(data, extensionRegistry); 500 } 501 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom(byte[] data) 502 throws com.google.protobuf.InvalidProtocolBufferException { 503 return PARSER.parseFrom(data); 504 } 505 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 506 byte[] data, 507 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 508 throws com.google.protobuf.InvalidProtocolBufferException { 509 return PARSER.parseFrom(data, extensionRegistry); 510 } 511 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom(java.io.InputStream input) 512 throws java.io.IOException { 513 return PARSER.parseFrom(input); 514 } 515 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 516 java.io.InputStream input, 517 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 518 throws java.io.IOException { 519 return PARSER.parseFrom(input, extensionRegistry); 520 } 521 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseDelimitedFrom(java.io.InputStream input) 522 throws java.io.IOException { 523 return PARSER.parseDelimitedFrom(input); 524 } 525 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseDelimitedFrom( 526 java.io.InputStream input, 527 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 528 throws java.io.IOException { 529 return PARSER.parseDelimitedFrom(input, extensionRegistry); 530 } 531 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 532 com.google.protobuf.CodedInputStream input) 533 throws java.io.IOException { 534 return PARSER.parseFrom(input); 535 } 536 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parseFrom( 537 com.google.protobuf.CodedInputStream input, 538 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 539 throws java.io.IOException { 540 return PARSER.parseFrom(input, extensionRegistry); 541 } 542 543 public static Builder newBuilder() { return Builder.create(); } 544 public Builder newBuilderForType() { return newBuilder(); } 545 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section prototype) { 546 return newBuilder().mergeFrom(prototype); 547 } 548 public Builder toBuilder() { return newBuilder(this); } 549 550 @java.lang.Override 551 protected Builder newBuilderForType( 552 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 553 Builder builder = new Builder(parent); 554 return builder; 555 } 556 /** 557 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary.Section} 558 * 559 * <pre> 560 * index for each section 561 * </pre> 562 */ 563 public static final class Builder extends 564 com.google.protobuf.GeneratedMessage.Builder<Builder> 565 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder { 566 public static final com.google.protobuf.Descriptors.Descriptor 567 getDescriptor() { 568 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 569 } 570 571 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 572 internalGetFieldAccessorTable() { 573 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable 574 .ensureFieldAccessorsInitialized( 575 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder.class); 576 } 577 578 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.newBuilder() 579 private Builder() { 580 maybeForceBuilderInitialization(); 581 } 582 583 private Builder( 584 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 585 super(parent); 586 maybeForceBuilderInitialization(); 587 } 588 private void maybeForceBuilderInitialization() { 589 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 590 } 591 } 592 private static Builder create() { 593 return new Builder(); 594 } 595 596 public Builder clear() { 597 super.clear(); 598 name_ = ""; 599 bitField0_ = (bitField0_ & ~0x00000001); 600 length_ = 0L; 601 bitField0_ = (bitField0_ & ~0x00000002); 602 offset_ = 0L; 603 bitField0_ = (bitField0_ & ~0x00000004); 604 return this; 605 } 606 607 public Builder clone() { 608 return create().mergeFrom(buildPartial()); 609 } 610 611 public com.google.protobuf.Descriptors.Descriptor 612 getDescriptorForType() { 613 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 614 } 615 616 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getDefaultInstanceForType() { 617 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance(); 618 } 619 620 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section build() { 621 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section result = buildPartial(); 622 if (!result.isInitialized()) { 623 throw newUninitializedMessageException(result); 624 } 625 return result; 626 } 627 628 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section buildPartial() { 629 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section(this); 630 int from_bitField0_ = bitField0_; 631 int to_bitField0_ = 0; 632 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 633 to_bitField0_ |= 0x00000001; 634 } 635 result.name_ = name_; 636 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 637 to_bitField0_ |= 0x00000002; 638 } 639 result.length_ = length_; 640 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 641 to_bitField0_ |= 0x00000004; 642 } 643 result.offset_ = offset_; 644 result.bitField0_ = to_bitField0_; 645 onBuilt(); 646 return result; 647 } 648 649 public Builder mergeFrom(com.google.protobuf.Message other) { 650 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section) { 651 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section)other); 652 } else { 653 super.mergeFrom(other); 654 return this; 655 } 656 } 657 658 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section other) { 659 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()) return this; 660 if (other.hasName()) { 661 bitField0_ |= 0x00000001; 662 name_ = other.name_; 663 onChanged(); 664 } 665 if (other.hasLength()) { 666 setLength(other.getLength()); 667 } 668 if (other.hasOffset()) { 669 setOffset(other.getOffset()); 670 } 671 this.mergeUnknownFields(other.getUnknownFields()); 672 return this; 673 } 674 675 public final boolean isInitialized() { 676 return true; 677 } 678 679 public Builder mergeFrom( 680 com.google.protobuf.CodedInputStream input, 681 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 682 throws java.io.IOException { 683 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section parsedMessage = null; 684 try { 685 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 686 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 687 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section) e.getUnfinishedMessage(); 688 throw e; 689 } finally { 690 if (parsedMessage != null) { 691 mergeFrom(parsedMessage); 692 } 693 } 694 return this; 695 } 696 private int bitField0_; 697 698 // optional string name = 1; 699 private java.lang.Object name_ = ""; 700 /** 701 * <code>optional string name = 1;</code> 702 */ 703 public boolean hasName() { 704 return ((bitField0_ & 0x00000001) == 0x00000001); 705 } 706 /** 707 * <code>optional string name = 1;</code> 708 */ 709 public java.lang.String getName() { 710 java.lang.Object ref = name_; 711 if (!(ref instanceof java.lang.String)) { 712 java.lang.String s = ((com.google.protobuf.ByteString) ref) 713 .toStringUtf8(); 714 name_ = s; 715 return s; 716 } else { 717 return (java.lang.String) ref; 718 } 719 } 720 /** 721 * <code>optional string name = 1;</code> 722 */ 723 public com.google.protobuf.ByteString 724 getNameBytes() { 725 java.lang.Object ref = name_; 726 if (ref instanceof String) { 727 com.google.protobuf.ByteString b = 728 com.google.protobuf.ByteString.copyFromUtf8( 729 (java.lang.String) ref); 730 name_ = b; 731 return b; 732 } else { 733 return (com.google.protobuf.ByteString) ref; 734 } 735 } 736 /** 737 * <code>optional string name = 1;</code> 738 */ 739 public Builder setName( 740 java.lang.String value) { 741 if (value == null) { 742 throw new NullPointerException(); 743 } 744 bitField0_ |= 0x00000001; 745 name_ = value; 746 onChanged(); 747 return this; 748 } 749 /** 750 * <code>optional string name = 1;</code> 751 */ 752 public Builder clearName() { 753 bitField0_ = (bitField0_ & ~0x00000001); 754 name_ = getDefaultInstance().getName(); 755 onChanged(); 756 return this; 757 } 758 /** 759 * <code>optional string name = 1;</code> 760 */ 761 public Builder setNameBytes( 762 com.google.protobuf.ByteString value) { 763 if (value == null) { 764 throw new NullPointerException(); 765 } 766 bitField0_ |= 0x00000001; 767 name_ = value; 768 onChanged(); 769 return this; 770 } 771 772 // optional uint64 length = 2; 773 private long length_ ; 774 /** 775 * <code>optional uint64 length = 2;</code> 776 */ 777 public boolean hasLength() { 778 return ((bitField0_ & 0x00000002) == 0x00000002); 779 } 780 /** 781 * <code>optional uint64 length = 2;</code> 782 */ 783 public long getLength() { 784 return length_; 785 } 786 /** 787 * <code>optional uint64 length = 2;</code> 788 */ 789 public Builder setLength(long value) { 790 bitField0_ |= 0x00000002; 791 length_ = value; 792 onChanged(); 793 return this; 794 } 795 /** 796 * <code>optional uint64 length = 2;</code> 797 */ 798 public Builder clearLength() { 799 bitField0_ = (bitField0_ & ~0x00000002); 800 length_ = 0L; 801 onChanged(); 802 return this; 803 } 804 805 // optional uint64 offset = 3; 806 private long offset_ ; 807 /** 808 * <code>optional uint64 offset = 3;</code> 809 */ 810 public boolean hasOffset() { 811 return ((bitField0_ & 0x00000004) == 0x00000004); 812 } 813 /** 814 * <code>optional uint64 offset = 3;</code> 815 */ 816 public long getOffset() { 817 return offset_; 818 } 819 /** 820 * <code>optional uint64 offset = 3;</code> 821 */ 822 public Builder setOffset(long value) { 823 bitField0_ |= 0x00000004; 824 offset_ = value; 825 onChanged(); 826 return this; 827 } 828 /** 829 * <code>optional uint64 offset = 3;</code> 830 */ 831 public Builder clearOffset() { 832 bitField0_ = (bitField0_ & ~0x00000004); 833 offset_ = 0L; 834 onChanged(); 835 return this; 836 } 837 838 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FileSummary.Section) 839 } 840 841 static { 842 defaultInstance = new Section(true); 843 defaultInstance.initFields(); 844 } 845 846 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FileSummary.Section) 847 } 848 849 private int bitField0_; 850 // required uint32 ondiskVersion = 1; 851 public static final int ONDISKVERSION_FIELD_NUMBER = 1; 852 private int ondiskVersion_; 853 /** 854 * <code>required uint32 ondiskVersion = 1;</code> 855 * 856 * <pre> 857 * The version of the above EBNF grammars. 858 * </pre> 859 */ 860 public boolean hasOndiskVersion() { 861 return ((bitField0_ & 0x00000001) == 0x00000001); 862 } 863 /** 864 * <code>required uint32 ondiskVersion = 1;</code> 865 * 866 * <pre> 867 * The version of the above EBNF grammars. 868 * </pre> 869 */ 870 public int getOndiskVersion() { 871 return ondiskVersion_; 872 } 873 874 // required uint32 layoutVersion = 2; 875 public static final int LAYOUTVERSION_FIELD_NUMBER = 2; 876 private int layoutVersion_; 877 /** 878 * <code>required uint32 layoutVersion = 2;</code> 879 * 880 * <pre> 881 * layoutVersion describes which features are available in the 882 * FSImage. 883 * </pre> 884 */ 885 public boolean hasLayoutVersion() { 886 return ((bitField0_ & 0x00000002) == 0x00000002); 887 } 888 /** 889 * <code>required uint32 layoutVersion = 2;</code> 890 * 891 * <pre> 892 * layoutVersion describes which features are available in the 893 * FSImage. 894 * </pre> 895 */ 896 public int getLayoutVersion() { 897 return layoutVersion_; 898 } 899 900 // optional string codec = 3; 901 public static final int CODEC_FIELD_NUMBER = 3; 902 private java.lang.Object codec_; 903 /** 904 * <code>optional string codec = 3;</code> 905 */ 906 public boolean hasCodec() { 907 return ((bitField0_ & 0x00000004) == 0x00000004); 908 } 909 /** 910 * <code>optional string codec = 3;</code> 911 */ 912 public java.lang.String getCodec() { 913 java.lang.Object ref = codec_; 914 if (ref instanceof java.lang.String) { 915 return (java.lang.String) ref; 916 } else { 917 com.google.protobuf.ByteString bs = 918 (com.google.protobuf.ByteString) ref; 919 java.lang.String s = bs.toStringUtf8(); 920 if (bs.isValidUtf8()) { 921 codec_ = s; 922 } 923 return s; 924 } 925 } 926 /** 927 * <code>optional string codec = 3;</code> 928 */ 929 public com.google.protobuf.ByteString 930 getCodecBytes() { 931 java.lang.Object ref = codec_; 932 if (ref instanceof java.lang.String) { 933 com.google.protobuf.ByteString b = 934 com.google.protobuf.ByteString.copyFromUtf8( 935 (java.lang.String) ref); 936 codec_ = b; 937 return b; 938 } else { 939 return (com.google.protobuf.ByteString) ref; 940 } 941 } 942 943 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 944 public static final int SECTIONS_FIELD_NUMBER = 4; 945 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> sections_; 946 /** 947 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 948 */ 949 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> getSectionsList() { 950 return sections_; 951 } 952 /** 953 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 954 */ 955 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 956 getSectionsOrBuilderList() { 957 return sections_; 958 } 959 /** 960 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 961 */ 962 public int getSectionsCount() { 963 return sections_.size(); 964 } 965 /** 966 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 967 */ 968 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index) { 969 return sections_.get(index); 970 } 971 /** 972 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 973 */ 974 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 975 int index) { 976 return sections_.get(index); 977 } 978 979 private void initFields() { 980 ondiskVersion_ = 0; 981 layoutVersion_ = 0; 982 codec_ = ""; 983 sections_ = java.util.Collections.emptyList(); 984 } 985 private byte memoizedIsInitialized = -1; 986 public final boolean isInitialized() { 987 byte isInitialized = memoizedIsInitialized; 988 if (isInitialized != -1) return isInitialized == 1; 989 990 if (!hasOndiskVersion()) { 991 memoizedIsInitialized = 0; 992 return false; 993 } 994 if (!hasLayoutVersion()) { 995 memoizedIsInitialized = 0; 996 return false; 997 } 998 memoizedIsInitialized = 1; 999 return true; 1000 } 1001 1002 public void writeTo(com.google.protobuf.CodedOutputStream output) 1003 throws java.io.IOException { 1004 getSerializedSize(); 1005 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1006 output.writeUInt32(1, ondiskVersion_); 1007 } 1008 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1009 output.writeUInt32(2, layoutVersion_); 1010 } 1011 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1012 output.writeBytes(3, getCodecBytes()); 1013 } 1014 for (int i = 0; i < sections_.size(); i++) { 1015 output.writeMessage(4, sections_.get(i)); 1016 } 1017 getUnknownFields().writeTo(output); 1018 } 1019 1020 private int memoizedSerializedSize = -1; 1021 public int getSerializedSize() { 1022 int size = memoizedSerializedSize; 1023 if (size != -1) return size; 1024 1025 size = 0; 1026 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1027 size += com.google.protobuf.CodedOutputStream 1028 .computeUInt32Size(1, ondiskVersion_); 1029 } 1030 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1031 size += com.google.protobuf.CodedOutputStream 1032 .computeUInt32Size(2, layoutVersion_); 1033 } 1034 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1035 size += com.google.protobuf.CodedOutputStream 1036 .computeBytesSize(3, getCodecBytes()); 1037 } 1038 for (int i = 0; i < sections_.size(); i++) { 1039 size += com.google.protobuf.CodedOutputStream 1040 .computeMessageSize(4, sections_.get(i)); 1041 } 1042 size += getUnknownFields().getSerializedSize(); 1043 memoizedSerializedSize = size; 1044 return size; 1045 } 1046 1047 private static final long serialVersionUID = 0L; 1048 @java.lang.Override 1049 protected java.lang.Object writeReplace() 1050 throws java.io.ObjectStreamException { 1051 return super.writeReplace(); 1052 } 1053 1054 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1055 com.google.protobuf.ByteString data) 1056 throws com.google.protobuf.InvalidProtocolBufferException { 1057 return PARSER.parseFrom(data); 1058 } 1059 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1060 com.google.protobuf.ByteString data, 1061 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1062 throws com.google.protobuf.InvalidProtocolBufferException { 1063 return PARSER.parseFrom(data, extensionRegistry); 1064 } 1065 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom(byte[] data) 1066 throws com.google.protobuf.InvalidProtocolBufferException { 1067 return PARSER.parseFrom(data); 1068 } 1069 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1070 byte[] data, 1071 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1072 throws com.google.protobuf.InvalidProtocolBufferException { 1073 return PARSER.parseFrom(data, extensionRegistry); 1074 } 1075 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom(java.io.InputStream input) 1076 throws java.io.IOException { 1077 return PARSER.parseFrom(input); 1078 } 1079 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1080 java.io.InputStream input, 1081 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1082 throws java.io.IOException { 1083 return PARSER.parseFrom(input, extensionRegistry); 1084 } 1085 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseDelimitedFrom(java.io.InputStream input) 1086 throws java.io.IOException { 1087 return PARSER.parseDelimitedFrom(input); 1088 } 1089 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseDelimitedFrom( 1090 java.io.InputStream input, 1091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1092 throws java.io.IOException { 1093 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1094 } 1095 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1096 com.google.protobuf.CodedInputStream input) 1097 throws java.io.IOException { 1098 return PARSER.parseFrom(input); 1099 } 1100 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parseFrom( 1101 com.google.protobuf.CodedInputStream input, 1102 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1103 throws java.io.IOException { 1104 return PARSER.parseFrom(input, extensionRegistry); 1105 } 1106 1107 public static Builder newBuilder() { return Builder.create(); } 1108 public Builder newBuilderForType() { return newBuilder(); } 1109 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary prototype) { 1110 return newBuilder().mergeFrom(prototype); 1111 } 1112 public Builder toBuilder() { return newBuilder(this); } 1113 1114 @java.lang.Override 1115 protected Builder newBuilderForType( 1116 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1117 Builder builder = new Builder(parent); 1118 return builder; 1119 } 1120 /** 1121 * Protobuf type {@code hadoop.hdfs.fsimage.FileSummary} 1122 */ 1123 public static final class Builder extends 1124 com.google.protobuf.GeneratedMessage.Builder<Builder> 1125 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummaryOrBuilder { 1126 public static final com.google.protobuf.Descriptors.Descriptor 1127 getDescriptor() { 1128 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 1129 } 1130 1131 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1132 internalGetFieldAccessorTable() { 1133 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable 1134 .ensureFieldAccessorsInitialized( 1135 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Builder.class); 1136 } 1137 1138 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.newBuilder() 1139 private Builder() { 1140 maybeForceBuilderInitialization(); 1141 } 1142 1143 private Builder( 1144 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1145 super(parent); 1146 maybeForceBuilderInitialization(); 1147 } 1148 private void maybeForceBuilderInitialization() { 1149 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1150 getSectionsFieldBuilder(); 1151 } 1152 } 1153 private static Builder create() { 1154 return new Builder(); 1155 } 1156 1157 public Builder clear() { 1158 super.clear(); 1159 ondiskVersion_ = 0; 1160 bitField0_ = (bitField0_ & ~0x00000001); 1161 layoutVersion_ = 0; 1162 bitField0_ = (bitField0_ & ~0x00000002); 1163 codec_ = ""; 1164 bitField0_ = (bitField0_ & ~0x00000004); 1165 if (sectionsBuilder_ == null) { 1166 sections_ = java.util.Collections.emptyList(); 1167 bitField0_ = (bitField0_ & ~0x00000008); 1168 } else { 1169 sectionsBuilder_.clear(); 1170 } 1171 return this; 1172 } 1173 1174 public Builder clone() { 1175 return create().mergeFrom(buildPartial()); 1176 } 1177 1178 public com.google.protobuf.Descriptors.Descriptor 1179 getDescriptorForType() { 1180 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 1181 } 1182 1183 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary getDefaultInstanceForType() { 1184 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.getDefaultInstance(); 1185 } 1186 1187 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary build() { 1188 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary result = buildPartial(); 1189 if (!result.isInitialized()) { 1190 throw newUninitializedMessageException(result); 1191 } 1192 return result; 1193 } 1194 1195 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary buildPartial() { 1196 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary(this); 1197 int from_bitField0_ = bitField0_; 1198 int to_bitField0_ = 0; 1199 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1200 to_bitField0_ |= 0x00000001; 1201 } 1202 result.ondiskVersion_ = ondiskVersion_; 1203 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 1204 to_bitField0_ |= 0x00000002; 1205 } 1206 result.layoutVersion_ = layoutVersion_; 1207 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 1208 to_bitField0_ |= 0x00000004; 1209 } 1210 result.codec_ = codec_; 1211 if (sectionsBuilder_ == null) { 1212 if (((bitField0_ & 0x00000008) == 0x00000008)) { 1213 sections_ = java.util.Collections.unmodifiableList(sections_); 1214 bitField0_ = (bitField0_ & ~0x00000008); 1215 } 1216 result.sections_ = sections_; 1217 } else { 1218 result.sections_ = sectionsBuilder_.build(); 1219 } 1220 result.bitField0_ = to_bitField0_; 1221 onBuilt(); 1222 return result; 1223 } 1224 1225 public Builder mergeFrom(com.google.protobuf.Message other) { 1226 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary) { 1227 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary)other); 1228 } else { 1229 super.mergeFrom(other); 1230 return this; 1231 } 1232 } 1233 1234 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary other) { 1235 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.getDefaultInstance()) return this; 1236 if (other.hasOndiskVersion()) { 1237 setOndiskVersion(other.getOndiskVersion()); 1238 } 1239 if (other.hasLayoutVersion()) { 1240 setLayoutVersion(other.getLayoutVersion()); 1241 } 1242 if (other.hasCodec()) { 1243 bitField0_ |= 0x00000004; 1244 codec_ = other.codec_; 1245 onChanged(); 1246 } 1247 if (sectionsBuilder_ == null) { 1248 if (!other.sections_.isEmpty()) { 1249 if (sections_.isEmpty()) { 1250 sections_ = other.sections_; 1251 bitField0_ = (bitField0_ & ~0x00000008); 1252 } else { 1253 ensureSectionsIsMutable(); 1254 sections_.addAll(other.sections_); 1255 } 1256 onChanged(); 1257 } 1258 } else { 1259 if (!other.sections_.isEmpty()) { 1260 if (sectionsBuilder_.isEmpty()) { 1261 sectionsBuilder_.dispose(); 1262 sectionsBuilder_ = null; 1263 sections_ = other.sections_; 1264 bitField0_ = (bitField0_ & ~0x00000008); 1265 sectionsBuilder_ = 1266 com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? 1267 getSectionsFieldBuilder() : null; 1268 } else { 1269 sectionsBuilder_.addAllMessages(other.sections_); 1270 } 1271 } 1272 } 1273 this.mergeUnknownFields(other.getUnknownFields()); 1274 return this; 1275 } 1276 1277 public final boolean isInitialized() { 1278 if (!hasOndiskVersion()) { 1279 1280 return false; 1281 } 1282 if (!hasLayoutVersion()) { 1283 1284 return false; 1285 } 1286 return true; 1287 } 1288 1289 public Builder mergeFrom( 1290 com.google.protobuf.CodedInputStream input, 1291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1292 throws java.io.IOException { 1293 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary parsedMessage = null; 1294 try { 1295 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1296 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1297 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary) e.getUnfinishedMessage(); 1298 throw e; 1299 } finally { 1300 if (parsedMessage != null) { 1301 mergeFrom(parsedMessage); 1302 } 1303 } 1304 return this; 1305 } 1306 private int bitField0_; 1307 1308 // required uint32 ondiskVersion = 1; 1309 private int ondiskVersion_ ; 1310 /** 1311 * <code>required uint32 ondiskVersion = 1;</code> 1312 * 1313 * <pre> 1314 * The version of the above EBNF grammars. 1315 * </pre> 1316 */ 1317 public boolean hasOndiskVersion() { 1318 return ((bitField0_ & 0x00000001) == 0x00000001); 1319 } 1320 /** 1321 * <code>required uint32 ondiskVersion = 1;</code> 1322 * 1323 * <pre> 1324 * The version of the above EBNF grammars. 1325 * </pre> 1326 */ 1327 public int getOndiskVersion() { 1328 return ondiskVersion_; 1329 } 1330 /** 1331 * <code>required uint32 ondiskVersion = 1;</code> 1332 * 1333 * <pre> 1334 * The version of the above EBNF grammars. 1335 * </pre> 1336 */ 1337 public Builder setOndiskVersion(int value) { 1338 bitField0_ |= 0x00000001; 1339 ondiskVersion_ = value; 1340 onChanged(); 1341 return this; 1342 } 1343 /** 1344 * <code>required uint32 ondiskVersion = 1;</code> 1345 * 1346 * <pre> 1347 * The version of the above EBNF grammars. 1348 * </pre> 1349 */ 1350 public Builder clearOndiskVersion() { 1351 bitField0_ = (bitField0_ & ~0x00000001); 1352 ondiskVersion_ = 0; 1353 onChanged(); 1354 return this; 1355 } 1356 1357 // required uint32 layoutVersion = 2; 1358 private int layoutVersion_ ; 1359 /** 1360 * <code>required uint32 layoutVersion = 2;</code> 1361 * 1362 * <pre> 1363 * layoutVersion describes which features are available in the 1364 * FSImage. 1365 * </pre> 1366 */ 1367 public boolean hasLayoutVersion() { 1368 return ((bitField0_ & 0x00000002) == 0x00000002); 1369 } 1370 /** 1371 * <code>required uint32 layoutVersion = 2;</code> 1372 * 1373 * <pre> 1374 * layoutVersion describes which features are available in the 1375 * FSImage. 1376 * </pre> 1377 */ 1378 public int getLayoutVersion() { 1379 return layoutVersion_; 1380 } 1381 /** 1382 * <code>required uint32 layoutVersion = 2;</code> 1383 * 1384 * <pre> 1385 * layoutVersion describes which features are available in the 1386 * FSImage. 1387 * </pre> 1388 */ 1389 public Builder setLayoutVersion(int value) { 1390 bitField0_ |= 0x00000002; 1391 layoutVersion_ = value; 1392 onChanged(); 1393 return this; 1394 } 1395 /** 1396 * <code>required uint32 layoutVersion = 2;</code> 1397 * 1398 * <pre> 1399 * layoutVersion describes which features are available in the 1400 * FSImage. 1401 * </pre> 1402 */ 1403 public Builder clearLayoutVersion() { 1404 bitField0_ = (bitField0_ & ~0x00000002); 1405 layoutVersion_ = 0; 1406 onChanged(); 1407 return this; 1408 } 1409 1410 // optional string codec = 3; 1411 private java.lang.Object codec_ = ""; 1412 /** 1413 * <code>optional string codec = 3;</code> 1414 */ 1415 public boolean hasCodec() { 1416 return ((bitField0_ & 0x00000004) == 0x00000004); 1417 } 1418 /** 1419 * <code>optional string codec = 3;</code> 1420 */ 1421 public java.lang.String getCodec() { 1422 java.lang.Object ref = codec_; 1423 if (!(ref instanceof java.lang.String)) { 1424 java.lang.String s = ((com.google.protobuf.ByteString) ref) 1425 .toStringUtf8(); 1426 codec_ = s; 1427 return s; 1428 } else { 1429 return (java.lang.String) ref; 1430 } 1431 } 1432 /** 1433 * <code>optional string codec = 3;</code> 1434 */ 1435 public com.google.protobuf.ByteString 1436 getCodecBytes() { 1437 java.lang.Object ref = codec_; 1438 if (ref instanceof String) { 1439 com.google.protobuf.ByteString b = 1440 com.google.protobuf.ByteString.copyFromUtf8( 1441 (java.lang.String) ref); 1442 codec_ = b; 1443 return b; 1444 } else { 1445 return (com.google.protobuf.ByteString) ref; 1446 } 1447 } 1448 /** 1449 * <code>optional string codec = 3;</code> 1450 */ 1451 public Builder setCodec( 1452 java.lang.String value) { 1453 if (value == null) { 1454 throw new NullPointerException(); 1455 } 1456 bitField0_ |= 0x00000004; 1457 codec_ = value; 1458 onChanged(); 1459 return this; 1460 } 1461 /** 1462 * <code>optional string codec = 3;</code> 1463 */ 1464 public Builder clearCodec() { 1465 bitField0_ = (bitField0_ & ~0x00000004); 1466 codec_ = getDefaultInstance().getCodec(); 1467 onChanged(); 1468 return this; 1469 } 1470 /** 1471 * <code>optional string codec = 3;</code> 1472 */ 1473 public Builder setCodecBytes( 1474 com.google.protobuf.ByteString value) { 1475 if (value == null) { 1476 throw new NullPointerException(); 1477 } 1478 bitField0_ |= 0x00000004; 1479 codec_ = value; 1480 onChanged(); 1481 return this; 1482 } 1483 1484 // repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4; 1485 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> sections_ = 1486 java.util.Collections.emptyList(); 1487 private void ensureSectionsIsMutable() { 1488 if (!((bitField0_ & 0x00000008) == 0x00000008)) { 1489 sections_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section>(sections_); 1490 bitField0_ |= 0x00000008; 1491 } 1492 } 1493 1494 private com.google.protobuf.RepeatedFieldBuilder< 1495 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> sectionsBuilder_; 1496 1497 /** 1498 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1499 */ 1500 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> getSectionsList() { 1501 if (sectionsBuilder_ == null) { 1502 return java.util.Collections.unmodifiableList(sections_); 1503 } else { 1504 return sectionsBuilder_.getMessageList(); 1505 } 1506 } 1507 /** 1508 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1509 */ 1510 public int getSectionsCount() { 1511 if (sectionsBuilder_ == null) { 1512 return sections_.size(); 1513 } else { 1514 return sectionsBuilder_.getCount(); 1515 } 1516 } 1517 /** 1518 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1519 */ 1520 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section getSections(int index) { 1521 if (sectionsBuilder_ == null) { 1522 return sections_.get(index); 1523 } else { 1524 return sectionsBuilder_.getMessage(index); 1525 } 1526 } 1527 /** 1528 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1529 */ 1530 public Builder setSections( 1531 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1532 if (sectionsBuilder_ == null) { 1533 if (value == null) { 1534 throw new NullPointerException(); 1535 } 1536 ensureSectionsIsMutable(); 1537 sections_.set(index, value); 1538 onChanged(); 1539 } else { 1540 sectionsBuilder_.setMessage(index, value); 1541 } 1542 return this; 1543 } 1544 /** 1545 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1546 */ 1547 public Builder setSections( 1548 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1549 if (sectionsBuilder_ == null) { 1550 ensureSectionsIsMutable(); 1551 sections_.set(index, builderForValue.build()); 1552 onChanged(); 1553 } else { 1554 sectionsBuilder_.setMessage(index, builderForValue.build()); 1555 } 1556 return this; 1557 } 1558 /** 1559 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1560 */ 1561 public Builder addSections(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1562 if (sectionsBuilder_ == null) { 1563 if (value == null) { 1564 throw new NullPointerException(); 1565 } 1566 ensureSectionsIsMutable(); 1567 sections_.add(value); 1568 onChanged(); 1569 } else { 1570 sectionsBuilder_.addMessage(value); 1571 } 1572 return this; 1573 } 1574 /** 1575 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1576 */ 1577 public Builder addSections( 1578 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section value) { 1579 if (sectionsBuilder_ == null) { 1580 if (value == null) { 1581 throw new NullPointerException(); 1582 } 1583 ensureSectionsIsMutable(); 1584 sections_.add(index, value); 1585 onChanged(); 1586 } else { 1587 sectionsBuilder_.addMessage(index, value); 1588 } 1589 return this; 1590 } 1591 /** 1592 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1593 */ 1594 public Builder addSections( 1595 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1596 if (sectionsBuilder_ == null) { 1597 ensureSectionsIsMutable(); 1598 sections_.add(builderForValue.build()); 1599 onChanged(); 1600 } else { 1601 sectionsBuilder_.addMessage(builderForValue.build()); 1602 } 1603 return this; 1604 } 1605 /** 1606 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1607 */ 1608 public Builder addSections( 1609 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder builderForValue) { 1610 if (sectionsBuilder_ == null) { 1611 ensureSectionsIsMutable(); 1612 sections_.add(index, builderForValue.build()); 1613 onChanged(); 1614 } else { 1615 sectionsBuilder_.addMessage(index, builderForValue.build()); 1616 } 1617 return this; 1618 } 1619 /** 1620 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1621 */ 1622 public Builder addAllSections( 1623 java.lang.Iterable<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section> values) { 1624 if (sectionsBuilder_ == null) { 1625 ensureSectionsIsMutable(); 1626 super.addAll(values, sections_); 1627 onChanged(); 1628 } else { 1629 sectionsBuilder_.addAllMessages(values); 1630 } 1631 return this; 1632 } 1633 /** 1634 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1635 */ 1636 public Builder clearSections() { 1637 if (sectionsBuilder_ == null) { 1638 sections_ = java.util.Collections.emptyList(); 1639 bitField0_ = (bitField0_ & ~0x00000008); 1640 onChanged(); 1641 } else { 1642 sectionsBuilder_.clear(); 1643 } 1644 return this; 1645 } 1646 /** 1647 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1648 */ 1649 public Builder removeSections(int index) { 1650 if (sectionsBuilder_ == null) { 1651 ensureSectionsIsMutable(); 1652 sections_.remove(index); 1653 onChanged(); 1654 } else { 1655 sectionsBuilder_.remove(index); 1656 } 1657 return this; 1658 } 1659 /** 1660 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1661 */ 1662 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder getSectionsBuilder( 1663 int index) { 1664 return getSectionsFieldBuilder().getBuilder(index); 1665 } 1666 /** 1667 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1668 */ 1669 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder getSectionsOrBuilder( 1670 int index) { 1671 if (sectionsBuilder_ == null) { 1672 return sections_.get(index); } else { 1673 return sectionsBuilder_.getMessageOrBuilder(index); 1674 } 1675 } 1676 /** 1677 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1678 */ 1679 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 1680 getSectionsOrBuilderList() { 1681 if (sectionsBuilder_ != null) { 1682 return sectionsBuilder_.getMessageOrBuilderList(); 1683 } else { 1684 return java.util.Collections.unmodifiableList(sections_); 1685 } 1686 } 1687 /** 1688 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1689 */ 1690 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder addSectionsBuilder() { 1691 return getSectionsFieldBuilder().addBuilder( 1692 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()); 1693 } 1694 /** 1695 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1696 */ 1697 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder addSectionsBuilder( 1698 int index) { 1699 return getSectionsFieldBuilder().addBuilder( 1700 index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.getDefaultInstance()); 1701 } 1702 /** 1703 * <code>repeated .hadoop.hdfs.fsimage.FileSummary.Section sections = 4;</code> 1704 */ 1705 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder> 1706 getSectionsBuilderList() { 1707 return getSectionsFieldBuilder().getBuilderList(); 1708 } 1709 private com.google.protobuf.RepeatedFieldBuilder< 1710 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder> 1711 getSectionsFieldBuilder() { 1712 if (sectionsBuilder_ == null) { 1713 sectionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< 1714 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.Section.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary.SectionOrBuilder>( 1715 sections_, 1716 ((bitField0_ & 0x00000008) == 0x00000008), 1717 getParentForChildren(), 1718 isClean()); 1719 sections_ = null; 1720 } 1721 return sectionsBuilder_; 1722 } 1723 1724 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FileSummary) 1725 } 1726 1727 static { 1728 defaultInstance = new FileSummary(true); 1729 defaultInstance.initFields(); 1730 } 1731 1732 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FileSummary) 1733 } 1734 1735 public interface NameSystemSectionOrBuilder 1736 extends com.google.protobuf.MessageOrBuilder { 1737 1738 // optional uint32 namespaceId = 1; 1739 /** 1740 * <code>optional uint32 namespaceId = 1;</code> 1741 */ 1742 boolean hasNamespaceId(); 1743 /** 1744 * <code>optional uint32 namespaceId = 1;</code> 1745 */ 1746 int getNamespaceId(); 1747 1748 // optional uint64 genstampV1 = 2; 1749 /** 1750 * <code>optional uint64 genstampV1 = 2;</code> 1751 */ 1752 boolean hasGenstampV1(); 1753 /** 1754 * <code>optional uint64 genstampV1 = 2;</code> 1755 */ 1756 long getGenstampV1(); 1757 1758 // optional uint64 genstampV2 = 3; 1759 /** 1760 * <code>optional uint64 genstampV2 = 3;</code> 1761 */ 1762 boolean hasGenstampV2(); 1763 /** 1764 * <code>optional uint64 genstampV2 = 3;</code> 1765 */ 1766 long getGenstampV2(); 1767 1768 // optional uint64 genstampV1Limit = 4; 1769 /** 1770 * <code>optional uint64 genstampV1Limit = 4;</code> 1771 */ 1772 boolean hasGenstampV1Limit(); 1773 /** 1774 * <code>optional uint64 genstampV1Limit = 4;</code> 1775 */ 1776 long getGenstampV1Limit(); 1777 1778 // optional uint64 lastAllocatedBlockId = 5; 1779 /** 1780 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 1781 */ 1782 boolean hasLastAllocatedBlockId(); 1783 /** 1784 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 1785 */ 1786 long getLastAllocatedBlockId(); 1787 1788 // optional uint64 transactionId = 6; 1789 /** 1790 * <code>optional uint64 transactionId = 6;</code> 1791 */ 1792 boolean hasTransactionId(); 1793 /** 1794 * <code>optional uint64 transactionId = 6;</code> 1795 */ 1796 long getTransactionId(); 1797 1798 // optional uint64 rollingUpgradeStartTime = 7; 1799 /** 1800 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 1801 */ 1802 boolean hasRollingUpgradeStartTime(); 1803 /** 1804 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 1805 */ 1806 long getRollingUpgradeStartTime(); 1807 } 1808 /** 1809 * Protobuf type {@code hadoop.hdfs.fsimage.NameSystemSection} 1810 * 1811 * <pre> 1812 ** 1813 * Name: NS_INFO 1814 * </pre> 1815 */ 1816 public static final class NameSystemSection extends 1817 com.google.protobuf.GeneratedMessage 1818 implements NameSystemSectionOrBuilder { 1819 // Use NameSystemSection.newBuilder() to construct. 1820 private NameSystemSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1821 super(builder); 1822 this.unknownFields = builder.getUnknownFields(); 1823 } 1824 private NameSystemSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1825 1826 private static final NameSystemSection defaultInstance; 1827 public static NameSystemSection getDefaultInstance() { 1828 return defaultInstance; 1829 } 1830 1831 public NameSystemSection getDefaultInstanceForType() { 1832 return defaultInstance; 1833 } 1834 1835 private final com.google.protobuf.UnknownFieldSet unknownFields; 1836 @java.lang.Override 1837 public final com.google.protobuf.UnknownFieldSet 1838 getUnknownFields() { 1839 return this.unknownFields; 1840 } 1841 private NameSystemSection( 1842 com.google.protobuf.CodedInputStream input, 1843 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1844 throws com.google.protobuf.InvalidProtocolBufferException { 1845 initFields(); 1846 int mutable_bitField0_ = 0; 1847 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1848 com.google.protobuf.UnknownFieldSet.newBuilder(); 1849 try { 1850 boolean done = false; 1851 while (!done) { 1852 int tag = input.readTag(); 1853 switch (tag) { 1854 case 0: 1855 done = true; 1856 break; 1857 default: { 1858 if (!parseUnknownField(input, unknownFields, 1859 extensionRegistry, tag)) { 1860 done = true; 1861 } 1862 break; 1863 } 1864 case 8: { 1865 bitField0_ |= 0x00000001; 1866 namespaceId_ = input.readUInt32(); 1867 break; 1868 } 1869 case 16: { 1870 bitField0_ |= 0x00000002; 1871 genstampV1_ = input.readUInt64(); 1872 break; 1873 } 1874 case 24: { 1875 bitField0_ |= 0x00000004; 1876 genstampV2_ = input.readUInt64(); 1877 break; 1878 } 1879 case 32: { 1880 bitField0_ |= 0x00000008; 1881 genstampV1Limit_ = input.readUInt64(); 1882 break; 1883 } 1884 case 40: { 1885 bitField0_ |= 0x00000010; 1886 lastAllocatedBlockId_ = input.readUInt64(); 1887 break; 1888 } 1889 case 48: { 1890 bitField0_ |= 0x00000020; 1891 transactionId_ = input.readUInt64(); 1892 break; 1893 } 1894 case 56: { 1895 bitField0_ |= 0x00000040; 1896 rollingUpgradeStartTime_ = input.readUInt64(); 1897 break; 1898 } 1899 } 1900 } 1901 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1902 throw e.setUnfinishedMessage(this); 1903 } catch (java.io.IOException e) { 1904 throw new com.google.protobuf.InvalidProtocolBufferException( 1905 e.getMessage()).setUnfinishedMessage(this); 1906 } finally { 1907 this.unknownFields = unknownFields.build(); 1908 makeExtensionsImmutable(); 1909 } 1910 } 1911 public static final com.google.protobuf.Descriptors.Descriptor 1912 getDescriptor() { 1913 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 1914 } 1915 1916 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1917 internalGetFieldAccessorTable() { 1918 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable 1919 .ensureFieldAccessorsInitialized( 1920 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.Builder.class); 1921 } 1922 1923 public static com.google.protobuf.Parser<NameSystemSection> PARSER = 1924 new com.google.protobuf.AbstractParser<NameSystemSection>() { 1925 public NameSystemSection parsePartialFrom( 1926 com.google.protobuf.CodedInputStream input, 1927 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1928 throws com.google.protobuf.InvalidProtocolBufferException { 1929 return new NameSystemSection(input, extensionRegistry); 1930 } 1931 }; 1932 1933 @java.lang.Override 1934 public com.google.protobuf.Parser<NameSystemSection> getParserForType() { 1935 return PARSER; 1936 } 1937 1938 private int bitField0_; 1939 // optional uint32 namespaceId = 1; 1940 public static final int NAMESPACEID_FIELD_NUMBER = 1; 1941 private int namespaceId_; 1942 /** 1943 * <code>optional uint32 namespaceId = 1;</code> 1944 */ 1945 public boolean hasNamespaceId() { 1946 return ((bitField0_ & 0x00000001) == 0x00000001); 1947 } 1948 /** 1949 * <code>optional uint32 namespaceId = 1;</code> 1950 */ 1951 public int getNamespaceId() { 1952 return namespaceId_; 1953 } 1954 1955 // optional uint64 genstampV1 = 2; 1956 public static final int GENSTAMPV1_FIELD_NUMBER = 2; 1957 private long genstampV1_; 1958 /** 1959 * <code>optional uint64 genstampV1 = 2;</code> 1960 */ 1961 public boolean hasGenstampV1() { 1962 return ((bitField0_ & 0x00000002) == 0x00000002); 1963 } 1964 /** 1965 * <code>optional uint64 genstampV1 = 2;</code> 1966 */ 1967 public long getGenstampV1() { 1968 return genstampV1_; 1969 } 1970 1971 // optional uint64 genstampV2 = 3; 1972 public static final int GENSTAMPV2_FIELD_NUMBER = 3; 1973 private long genstampV2_; 1974 /** 1975 * <code>optional uint64 genstampV2 = 3;</code> 1976 */ 1977 public boolean hasGenstampV2() { 1978 return ((bitField0_ & 0x00000004) == 0x00000004); 1979 } 1980 /** 1981 * <code>optional uint64 genstampV2 = 3;</code> 1982 */ 1983 public long getGenstampV2() { 1984 return genstampV2_; 1985 } 1986 1987 // optional uint64 genstampV1Limit = 4; 1988 public static final int GENSTAMPV1LIMIT_FIELD_NUMBER = 4; 1989 private long genstampV1Limit_; 1990 /** 1991 * <code>optional uint64 genstampV1Limit = 4;</code> 1992 */ 1993 public boolean hasGenstampV1Limit() { 1994 return ((bitField0_ & 0x00000008) == 0x00000008); 1995 } 1996 /** 1997 * <code>optional uint64 genstampV1Limit = 4;</code> 1998 */ 1999 public long getGenstampV1Limit() { 2000 return genstampV1Limit_; 2001 } 2002 2003 // optional uint64 lastAllocatedBlockId = 5; 2004 public static final int LASTALLOCATEDBLOCKID_FIELD_NUMBER = 5; 2005 private long lastAllocatedBlockId_; 2006 /** 2007 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2008 */ 2009 public boolean hasLastAllocatedBlockId() { 2010 return ((bitField0_ & 0x00000010) == 0x00000010); 2011 } 2012 /** 2013 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2014 */ 2015 public long getLastAllocatedBlockId() { 2016 return lastAllocatedBlockId_; 2017 } 2018 2019 // optional uint64 transactionId = 6; 2020 public static final int TRANSACTIONID_FIELD_NUMBER = 6; 2021 private long transactionId_; 2022 /** 2023 * <code>optional uint64 transactionId = 6;</code> 2024 */ 2025 public boolean hasTransactionId() { 2026 return ((bitField0_ & 0x00000020) == 0x00000020); 2027 } 2028 /** 2029 * <code>optional uint64 transactionId = 6;</code> 2030 */ 2031 public long getTransactionId() { 2032 return transactionId_; 2033 } 2034 2035 // optional uint64 rollingUpgradeStartTime = 7; 2036 public static final int ROLLINGUPGRADESTARTTIME_FIELD_NUMBER = 7; 2037 private long rollingUpgradeStartTime_; 2038 /** 2039 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2040 */ 2041 public boolean hasRollingUpgradeStartTime() { 2042 return ((bitField0_ & 0x00000040) == 0x00000040); 2043 } 2044 /** 2045 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2046 */ 2047 public long getRollingUpgradeStartTime() { 2048 return rollingUpgradeStartTime_; 2049 } 2050 2051 private void initFields() { 2052 namespaceId_ = 0; 2053 genstampV1_ = 0L; 2054 genstampV2_ = 0L; 2055 genstampV1Limit_ = 0L; 2056 lastAllocatedBlockId_ = 0L; 2057 transactionId_ = 0L; 2058 rollingUpgradeStartTime_ = 0L; 2059 } 2060 private byte memoizedIsInitialized = -1; 2061 public final boolean isInitialized() { 2062 byte isInitialized = memoizedIsInitialized; 2063 if (isInitialized != -1) return isInitialized == 1; 2064 2065 memoizedIsInitialized = 1; 2066 return true; 2067 } 2068 2069 public void writeTo(com.google.protobuf.CodedOutputStream output) 2070 throws java.io.IOException { 2071 getSerializedSize(); 2072 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2073 output.writeUInt32(1, namespaceId_); 2074 } 2075 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2076 output.writeUInt64(2, genstampV1_); 2077 } 2078 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2079 output.writeUInt64(3, genstampV2_); 2080 } 2081 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2082 output.writeUInt64(4, genstampV1Limit_); 2083 } 2084 if (((bitField0_ & 0x00000010) == 0x00000010)) { 2085 output.writeUInt64(5, lastAllocatedBlockId_); 2086 } 2087 if (((bitField0_ & 0x00000020) == 0x00000020)) { 2088 output.writeUInt64(6, transactionId_); 2089 } 2090 if (((bitField0_ & 0x00000040) == 0x00000040)) { 2091 output.writeUInt64(7, rollingUpgradeStartTime_); 2092 } 2093 getUnknownFields().writeTo(output); 2094 } 2095 2096 private int memoizedSerializedSize = -1; 2097 public int getSerializedSize() { 2098 int size = memoizedSerializedSize; 2099 if (size != -1) return size; 2100 2101 size = 0; 2102 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2103 size += com.google.protobuf.CodedOutputStream 2104 .computeUInt32Size(1, namespaceId_); 2105 } 2106 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2107 size += com.google.protobuf.CodedOutputStream 2108 .computeUInt64Size(2, genstampV1_); 2109 } 2110 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2111 size += com.google.protobuf.CodedOutputStream 2112 .computeUInt64Size(3, genstampV2_); 2113 } 2114 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2115 size += com.google.protobuf.CodedOutputStream 2116 .computeUInt64Size(4, genstampV1Limit_); 2117 } 2118 if (((bitField0_ & 0x00000010) == 0x00000010)) { 2119 size += com.google.protobuf.CodedOutputStream 2120 .computeUInt64Size(5, lastAllocatedBlockId_); 2121 } 2122 if (((bitField0_ & 0x00000020) == 0x00000020)) { 2123 size += com.google.protobuf.CodedOutputStream 2124 .computeUInt64Size(6, transactionId_); 2125 } 2126 if (((bitField0_ & 0x00000040) == 0x00000040)) { 2127 size += com.google.protobuf.CodedOutputStream 2128 .computeUInt64Size(7, rollingUpgradeStartTime_); 2129 } 2130 size += getUnknownFields().getSerializedSize(); 2131 memoizedSerializedSize = size; 2132 return size; 2133 } 2134 2135 private static final long serialVersionUID = 0L; 2136 @java.lang.Override 2137 protected java.lang.Object writeReplace() 2138 throws java.io.ObjectStreamException { 2139 return super.writeReplace(); 2140 } 2141 2142 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2143 com.google.protobuf.ByteString data) 2144 throws com.google.protobuf.InvalidProtocolBufferException { 2145 return PARSER.parseFrom(data); 2146 } 2147 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2148 com.google.protobuf.ByteString data, 2149 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2150 throws com.google.protobuf.InvalidProtocolBufferException { 2151 return PARSER.parseFrom(data, extensionRegistry); 2152 } 2153 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom(byte[] data) 2154 throws com.google.protobuf.InvalidProtocolBufferException { 2155 return PARSER.parseFrom(data); 2156 } 2157 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2158 byte[] data, 2159 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2160 throws com.google.protobuf.InvalidProtocolBufferException { 2161 return PARSER.parseFrom(data, extensionRegistry); 2162 } 2163 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom(java.io.InputStream input) 2164 throws java.io.IOException { 2165 return PARSER.parseFrom(input); 2166 } 2167 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2168 java.io.InputStream input, 2169 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2170 throws java.io.IOException { 2171 return PARSER.parseFrom(input, extensionRegistry); 2172 } 2173 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseDelimitedFrom(java.io.InputStream input) 2174 throws java.io.IOException { 2175 return PARSER.parseDelimitedFrom(input); 2176 } 2177 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseDelimitedFrom( 2178 java.io.InputStream input, 2179 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2180 throws java.io.IOException { 2181 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2182 } 2183 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2184 com.google.protobuf.CodedInputStream input) 2185 throws java.io.IOException { 2186 return PARSER.parseFrom(input); 2187 } 2188 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parseFrom( 2189 com.google.protobuf.CodedInputStream input, 2190 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2191 throws java.io.IOException { 2192 return PARSER.parseFrom(input, extensionRegistry); 2193 } 2194 2195 public static Builder newBuilder() { return Builder.create(); } 2196 public Builder newBuilderForType() { return newBuilder(); } 2197 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection prototype) { 2198 return newBuilder().mergeFrom(prototype); 2199 } 2200 public Builder toBuilder() { return newBuilder(this); } 2201 2202 @java.lang.Override 2203 protected Builder newBuilderForType( 2204 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2205 Builder builder = new Builder(parent); 2206 return builder; 2207 } 2208 /** 2209 * Protobuf type {@code hadoop.hdfs.fsimage.NameSystemSection} 2210 * 2211 * <pre> 2212 ** 2213 * Name: NS_INFO 2214 * </pre> 2215 */ 2216 public static final class Builder extends 2217 com.google.protobuf.GeneratedMessage.Builder<Builder> 2218 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSectionOrBuilder { 2219 public static final com.google.protobuf.Descriptors.Descriptor 2220 getDescriptor() { 2221 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 2222 } 2223 2224 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2225 internalGetFieldAccessorTable() { 2226 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable 2227 .ensureFieldAccessorsInitialized( 2228 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.Builder.class); 2229 } 2230 2231 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.newBuilder() 2232 private Builder() { 2233 maybeForceBuilderInitialization(); 2234 } 2235 2236 private Builder( 2237 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2238 super(parent); 2239 maybeForceBuilderInitialization(); 2240 } 2241 private void maybeForceBuilderInitialization() { 2242 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2243 } 2244 } 2245 private static Builder create() { 2246 return new Builder(); 2247 } 2248 2249 public Builder clear() { 2250 super.clear(); 2251 namespaceId_ = 0; 2252 bitField0_ = (bitField0_ & ~0x00000001); 2253 genstampV1_ = 0L; 2254 bitField0_ = (bitField0_ & ~0x00000002); 2255 genstampV2_ = 0L; 2256 bitField0_ = (bitField0_ & ~0x00000004); 2257 genstampV1Limit_ = 0L; 2258 bitField0_ = (bitField0_ & ~0x00000008); 2259 lastAllocatedBlockId_ = 0L; 2260 bitField0_ = (bitField0_ & ~0x00000010); 2261 transactionId_ = 0L; 2262 bitField0_ = (bitField0_ & ~0x00000020); 2263 rollingUpgradeStartTime_ = 0L; 2264 bitField0_ = (bitField0_ & ~0x00000040); 2265 return this; 2266 } 2267 2268 public Builder clone() { 2269 return create().mergeFrom(buildPartial()); 2270 } 2271 2272 public com.google.protobuf.Descriptors.Descriptor 2273 getDescriptorForType() { 2274 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 2275 } 2276 2277 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection getDefaultInstanceForType() { 2278 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.getDefaultInstance(); 2279 } 2280 2281 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection build() { 2282 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection result = buildPartial(); 2283 if (!result.isInitialized()) { 2284 throw newUninitializedMessageException(result); 2285 } 2286 return result; 2287 } 2288 2289 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection buildPartial() { 2290 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection(this); 2291 int from_bitField0_ = bitField0_; 2292 int to_bitField0_ = 0; 2293 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 2294 to_bitField0_ |= 0x00000001; 2295 } 2296 result.namespaceId_ = namespaceId_; 2297 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 2298 to_bitField0_ |= 0x00000002; 2299 } 2300 result.genstampV1_ = genstampV1_; 2301 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 2302 to_bitField0_ |= 0x00000004; 2303 } 2304 result.genstampV2_ = genstampV2_; 2305 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 2306 to_bitField0_ |= 0x00000008; 2307 } 2308 result.genstampV1Limit_ = genstampV1Limit_; 2309 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 2310 to_bitField0_ |= 0x00000010; 2311 } 2312 result.lastAllocatedBlockId_ = lastAllocatedBlockId_; 2313 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 2314 to_bitField0_ |= 0x00000020; 2315 } 2316 result.transactionId_ = transactionId_; 2317 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 2318 to_bitField0_ |= 0x00000040; 2319 } 2320 result.rollingUpgradeStartTime_ = rollingUpgradeStartTime_; 2321 result.bitField0_ = to_bitField0_; 2322 onBuilt(); 2323 return result; 2324 } 2325 2326 public Builder mergeFrom(com.google.protobuf.Message other) { 2327 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection) { 2328 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection)other); 2329 } else { 2330 super.mergeFrom(other); 2331 return this; 2332 } 2333 } 2334 2335 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection other) { 2336 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection.getDefaultInstance()) return this; 2337 if (other.hasNamespaceId()) { 2338 setNamespaceId(other.getNamespaceId()); 2339 } 2340 if (other.hasGenstampV1()) { 2341 setGenstampV1(other.getGenstampV1()); 2342 } 2343 if (other.hasGenstampV2()) { 2344 setGenstampV2(other.getGenstampV2()); 2345 } 2346 if (other.hasGenstampV1Limit()) { 2347 setGenstampV1Limit(other.getGenstampV1Limit()); 2348 } 2349 if (other.hasLastAllocatedBlockId()) { 2350 setLastAllocatedBlockId(other.getLastAllocatedBlockId()); 2351 } 2352 if (other.hasTransactionId()) { 2353 setTransactionId(other.getTransactionId()); 2354 } 2355 if (other.hasRollingUpgradeStartTime()) { 2356 setRollingUpgradeStartTime(other.getRollingUpgradeStartTime()); 2357 } 2358 this.mergeUnknownFields(other.getUnknownFields()); 2359 return this; 2360 } 2361 2362 public final boolean isInitialized() { 2363 return true; 2364 } 2365 2366 public Builder mergeFrom( 2367 com.google.protobuf.CodedInputStream input, 2368 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2369 throws java.io.IOException { 2370 org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection parsedMessage = null; 2371 try { 2372 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2373 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2374 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection) e.getUnfinishedMessage(); 2375 throw e; 2376 } finally { 2377 if (parsedMessage != null) { 2378 mergeFrom(parsedMessage); 2379 } 2380 } 2381 return this; 2382 } 2383 private int bitField0_; 2384 2385 // optional uint32 namespaceId = 1; 2386 private int namespaceId_ ; 2387 /** 2388 * <code>optional uint32 namespaceId = 1;</code> 2389 */ 2390 public boolean hasNamespaceId() { 2391 return ((bitField0_ & 0x00000001) == 0x00000001); 2392 } 2393 /** 2394 * <code>optional uint32 namespaceId = 1;</code> 2395 */ 2396 public int getNamespaceId() { 2397 return namespaceId_; 2398 } 2399 /** 2400 * <code>optional uint32 namespaceId = 1;</code> 2401 */ 2402 public Builder setNamespaceId(int value) { 2403 bitField0_ |= 0x00000001; 2404 namespaceId_ = value; 2405 onChanged(); 2406 return this; 2407 } 2408 /** 2409 * <code>optional uint32 namespaceId = 1;</code> 2410 */ 2411 public Builder clearNamespaceId() { 2412 bitField0_ = (bitField0_ & ~0x00000001); 2413 namespaceId_ = 0; 2414 onChanged(); 2415 return this; 2416 } 2417 2418 // optional uint64 genstampV1 = 2; 2419 private long genstampV1_ ; 2420 /** 2421 * <code>optional uint64 genstampV1 = 2;</code> 2422 */ 2423 public boolean hasGenstampV1() { 2424 return ((bitField0_ & 0x00000002) == 0x00000002); 2425 } 2426 /** 2427 * <code>optional uint64 genstampV1 = 2;</code> 2428 */ 2429 public long getGenstampV1() { 2430 return genstampV1_; 2431 } 2432 /** 2433 * <code>optional uint64 genstampV1 = 2;</code> 2434 */ 2435 public Builder setGenstampV1(long value) { 2436 bitField0_ |= 0x00000002; 2437 genstampV1_ = value; 2438 onChanged(); 2439 return this; 2440 } 2441 /** 2442 * <code>optional uint64 genstampV1 = 2;</code> 2443 */ 2444 public Builder clearGenstampV1() { 2445 bitField0_ = (bitField0_ & ~0x00000002); 2446 genstampV1_ = 0L; 2447 onChanged(); 2448 return this; 2449 } 2450 2451 // optional uint64 genstampV2 = 3; 2452 private long genstampV2_ ; 2453 /** 2454 * <code>optional uint64 genstampV2 = 3;</code> 2455 */ 2456 public boolean hasGenstampV2() { 2457 return ((bitField0_ & 0x00000004) == 0x00000004); 2458 } 2459 /** 2460 * <code>optional uint64 genstampV2 = 3;</code> 2461 */ 2462 public long getGenstampV2() { 2463 return genstampV2_; 2464 } 2465 /** 2466 * <code>optional uint64 genstampV2 = 3;</code> 2467 */ 2468 public Builder setGenstampV2(long value) { 2469 bitField0_ |= 0x00000004; 2470 genstampV2_ = value; 2471 onChanged(); 2472 return this; 2473 } 2474 /** 2475 * <code>optional uint64 genstampV2 = 3;</code> 2476 */ 2477 public Builder clearGenstampV2() { 2478 bitField0_ = (bitField0_ & ~0x00000004); 2479 genstampV2_ = 0L; 2480 onChanged(); 2481 return this; 2482 } 2483 2484 // optional uint64 genstampV1Limit = 4; 2485 private long genstampV1Limit_ ; 2486 /** 2487 * <code>optional uint64 genstampV1Limit = 4;</code> 2488 */ 2489 public boolean hasGenstampV1Limit() { 2490 return ((bitField0_ & 0x00000008) == 0x00000008); 2491 } 2492 /** 2493 * <code>optional uint64 genstampV1Limit = 4;</code> 2494 */ 2495 public long getGenstampV1Limit() { 2496 return genstampV1Limit_; 2497 } 2498 /** 2499 * <code>optional uint64 genstampV1Limit = 4;</code> 2500 */ 2501 public Builder setGenstampV1Limit(long value) { 2502 bitField0_ |= 0x00000008; 2503 genstampV1Limit_ = value; 2504 onChanged(); 2505 return this; 2506 } 2507 /** 2508 * <code>optional uint64 genstampV1Limit = 4;</code> 2509 */ 2510 public Builder clearGenstampV1Limit() { 2511 bitField0_ = (bitField0_ & ~0x00000008); 2512 genstampV1Limit_ = 0L; 2513 onChanged(); 2514 return this; 2515 } 2516 2517 // optional uint64 lastAllocatedBlockId = 5; 2518 private long lastAllocatedBlockId_ ; 2519 /** 2520 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2521 */ 2522 public boolean hasLastAllocatedBlockId() { 2523 return ((bitField0_ & 0x00000010) == 0x00000010); 2524 } 2525 /** 2526 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2527 */ 2528 public long getLastAllocatedBlockId() { 2529 return lastAllocatedBlockId_; 2530 } 2531 /** 2532 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2533 */ 2534 public Builder setLastAllocatedBlockId(long value) { 2535 bitField0_ |= 0x00000010; 2536 lastAllocatedBlockId_ = value; 2537 onChanged(); 2538 return this; 2539 } 2540 /** 2541 * <code>optional uint64 lastAllocatedBlockId = 5;</code> 2542 */ 2543 public Builder clearLastAllocatedBlockId() { 2544 bitField0_ = (bitField0_ & ~0x00000010); 2545 lastAllocatedBlockId_ = 0L; 2546 onChanged(); 2547 return this; 2548 } 2549 2550 // optional uint64 transactionId = 6; 2551 private long transactionId_ ; 2552 /** 2553 * <code>optional uint64 transactionId = 6;</code> 2554 */ 2555 public boolean hasTransactionId() { 2556 return ((bitField0_ & 0x00000020) == 0x00000020); 2557 } 2558 /** 2559 * <code>optional uint64 transactionId = 6;</code> 2560 */ 2561 public long getTransactionId() { 2562 return transactionId_; 2563 } 2564 /** 2565 * <code>optional uint64 transactionId = 6;</code> 2566 */ 2567 public Builder setTransactionId(long value) { 2568 bitField0_ |= 0x00000020; 2569 transactionId_ = value; 2570 onChanged(); 2571 return this; 2572 } 2573 /** 2574 * <code>optional uint64 transactionId = 6;</code> 2575 */ 2576 public Builder clearTransactionId() { 2577 bitField0_ = (bitField0_ & ~0x00000020); 2578 transactionId_ = 0L; 2579 onChanged(); 2580 return this; 2581 } 2582 2583 // optional uint64 rollingUpgradeStartTime = 7; 2584 private long rollingUpgradeStartTime_ ; 2585 /** 2586 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2587 */ 2588 public boolean hasRollingUpgradeStartTime() { 2589 return ((bitField0_ & 0x00000040) == 0x00000040); 2590 } 2591 /** 2592 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2593 */ 2594 public long getRollingUpgradeStartTime() { 2595 return rollingUpgradeStartTime_; 2596 } 2597 /** 2598 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2599 */ 2600 public Builder setRollingUpgradeStartTime(long value) { 2601 bitField0_ |= 0x00000040; 2602 rollingUpgradeStartTime_ = value; 2603 onChanged(); 2604 return this; 2605 } 2606 /** 2607 * <code>optional uint64 rollingUpgradeStartTime = 7;</code> 2608 */ 2609 public Builder clearRollingUpgradeStartTime() { 2610 bitField0_ = (bitField0_ & ~0x00000040); 2611 rollingUpgradeStartTime_ = 0L; 2612 onChanged(); 2613 return this; 2614 } 2615 2616 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.NameSystemSection) 2617 } 2618 2619 static { 2620 defaultInstance = new NameSystemSection(true); 2621 defaultInstance.initFields(); 2622 } 2623 2624 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.NameSystemSection) 2625 } 2626 2627 public interface INodeSectionOrBuilder 2628 extends com.google.protobuf.MessageOrBuilder { 2629 2630 // optional uint64 lastInodeId = 1; 2631 /** 2632 * <code>optional uint64 lastInodeId = 1;</code> 2633 */ 2634 boolean hasLastInodeId(); 2635 /** 2636 * <code>optional uint64 lastInodeId = 1;</code> 2637 */ 2638 long getLastInodeId(); 2639 2640 // optional uint64 numInodes = 2; 2641 /** 2642 * <code>optional uint64 numInodes = 2;</code> 2643 * 2644 * <pre> 2645 * repeated INodes.. 2646 * </pre> 2647 */ 2648 boolean hasNumInodes(); 2649 /** 2650 * <code>optional uint64 numInodes = 2;</code> 2651 * 2652 * <pre> 2653 * repeated INodes.. 2654 * </pre> 2655 */ 2656 long getNumInodes(); 2657 } 2658 /** 2659 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection} 2660 * 2661 * <pre> 2662 ** 2663 * Permission is serialized as a 64-bit long. [0:24):[25:48):[48:64) (in Big Endian). 2664 * The first and the second parts are the string ids of the user and 2665 * group name, and the last 16 bits are the permission bits. 2666 * 2667 * Name: INODE 2668 * </pre> 2669 */ 2670 public static final class INodeSection extends 2671 com.google.protobuf.GeneratedMessage 2672 implements INodeSectionOrBuilder { 2673 // Use INodeSection.newBuilder() to construct. 2674 private INodeSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2675 super(builder); 2676 this.unknownFields = builder.getUnknownFields(); 2677 } 2678 private INodeSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2679 2680 private static final INodeSection defaultInstance; 2681 public static INodeSection getDefaultInstance() { 2682 return defaultInstance; 2683 } 2684 2685 public INodeSection getDefaultInstanceForType() { 2686 return defaultInstance; 2687 } 2688 2689 private final com.google.protobuf.UnknownFieldSet unknownFields; 2690 @java.lang.Override 2691 public final com.google.protobuf.UnknownFieldSet 2692 getUnknownFields() { 2693 return this.unknownFields; 2694 } 2695 private INodeSection( 2696 com.google.protobuf.CodedInputStream input, 2697 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2698 throws com.google.protobuf.InvalidProtocolBufferException { 2699 initFields(); 2700 int mutable_bitField0_ = 0; 2701 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2702 com.google.protobuf.UnknownFieldSet.newBuilder(); 2703 try { 2704 boolean done = false; 2705 while (!done) { 2706 int tag = input.readTag(); 2707 switch (tag) { 2708 case 0: 2709 done = true; 2710 break; 2711 default: { 2712 if (!parseUnknownField(input, unknownFields, 2713 extensionRegistry, tag)) { 2714 done = true; 2715 } 2716 break; 2717 } 2718 case 8: { 2719 bitField0_ |= 0x00000001; 2720 lastInodeId_ = input.readUInt64(); 2721 break; 2722 } 2723 case 16: { 2724 bitField0_ |= 0x00000002; 2725 numInodes_ = input.readUInt64(); 2726 break; 2727 } 2728 } 2729 } 2730 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2731 throw e.setUnfinishedMessage(this); 2732 } catch (java.io.IOException e) { 2733 throw new com.google.protobuf.InvalidProtocolBufferException( 2734 e.getMessage()).setUnfinishedMessage(this); 2735 } finally { 2736 this.unknownFields = unknownFields.build(); 2737 makeExtensionsImmutable(); 2738 } 2739 } 2740 public static final com.google.protobuf.Descriptors.Descriptor 2741 getDescriptor() { 2742 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 2743 } 2744 2745 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2746 internalGetFieldAccessorTable() { 2747 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable 2748 .ensureFieldAccessorsInitialized( 2749 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.Builder.class); 2750 } 2751 2752 public static com.google.protobuf.Parser<INodeSection> PARSER = 2753 new com.google.protobuf.AbstractParser<INodeSection>() { 2754 public INodeSection parsePartialFrom( 2755 com.google.protobuf.CodedInputStream input, 2756 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2757 throws com.google.protobuf.InvalidProtocolBufferException { 2758 return new INodeSection(input, extensionRegistry); 2759 } 2760 }; 2761 2762 @java.lang.Override 2763 public com.google.protobuf.Parser<INodeSection> getParserForType() { 2764 return PARSER; 2765 } 2766 2767 public interface FileUnderConstructionFeatureOrBuilder 2768 extends com.google.protobuf.MessageOrBuilder { 2769 2770 // optional string clientName = 1; 2771 /** 2772 * <code>optional string clientName = 1;</code> 2773 */ 2774 boolean hasClientName(); 2775 /** 2776 * <code>optional string clientName = 1;</code> 2777 */ 2778 java.lang.String getClientName(); 2779 /** 2780 * <code>optional string clientName = 1;</code> 2781 */ 2782 com.google.protobuf.ByteString 2783 getClientNameBytes(); 2784 2785 // optional string clientMachine = 2; 2786 /** 2787 * <code>optional string clientMachine = 2;</code> 2788 */ 2789 boolean hasClientMachine(); 2790 /** 2791 * <code>optional string clientMachine = 2;</code> 2792 */ 2793 java.lang.String getClientMachine(); 2794 /** 2795 * <code>optional string clientMachine = 2;</code> 2796 */ 2797 com.google.protobuf.ByteString 2798 getClientMachineBytes(); 2799 } 2800 /** 2801 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature} 2802 * 2803 * <pre> 2804 ** 2805 * under-construction feature for INodeFile 2806 * </pre> 2807 */ 2808 public static final class FileUnderConstructionFeature extends 2809 com.google.protobuf.GeneratedMessage 2810 implements FileUnderConstructionFeatureOrBuilder { 2811 // Use FileUnderConstructionFeature.newBuilder() to construct. 2812 private FileUnderConstructionFeature(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2813 super(builder); 2814 this.unknownFields = builder.getUnknownFields(); 2815 } 2816 private FileUnderConstructionFeature(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2817 2818 private static final FileUnderConstructionFeature defaultInstance; 2819 public static FileUnderConstructionFeature getDefaultInstance() { 2820 return defaultInstance; 2821 } 2822 2823 public FileUnderConstructionFeature getDefaultInstanceForType() { 2824 return defaultInstance; 2825 } 2826 2827 private final com.google.protobuf.UnknownFieldSet unknownFields; 2828 @java.lang.Override 2829 public final com.google.protobuf.UnknownFieldSet 2830 getUnknownFields() { 2831 return this.unknownFields; 2832 } 2833 private FileUnderConstructionFeature( 2834 com.google.protobuf.CodedInputStream input, 2835 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2836 throws com.google.protobuf.InvalidProtocolBufferException { 2837 initFields(); 2838 int mutable_bitField0_ = 0; 2839 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2840 com.google.protobuf.UnknownFieldSet.newBuilder(); 2841 try { 2842 boolean done = false; 2843 while (!done) { 2844 int tag = input.readTag(); 2845 switch (tag) { 2846 case 0: 2847 done = true; 2848 break; 2849 default: { 2850 if (!parseUnknownField(input, unknownFields, 2851 extensionRegistry, tag)) { 2852 done = true; 2853 } 2854 break; 2855 } 2856 case 10: { 2857 bitField0_ |= 0x00000001; 2858 clientName_ = input.readBytes(); 2859 break; 2860 } 2861 case 18: { 2862 bitField0_ |= 0x00000002; 2863 clientMachine_ = input.readBytes(); 2864 break; 2865 } 2866 } 2867 } 2868 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2869 throw e.setUnfinishedMessage(this); 2870 } catch (java.io.IOException e) { 2871 throw new com.google.protobuf.InvalidProtocolBufferException( 2872 e.getMessage()).setUnfinishedMessage(this); 2873 } finally { 2874 this.unknownFields = unknownFields.build(); 2875 makeExtensionsImmutable(); 2876 } 2877 } 2878 public static final com.google.protobuf.Descriptors.Descriptor 2879 getDescriptor() { 2880 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 2881 } 2882 2883 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2884 internalGetFieldAccessorTable() { 2885 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable 2886 .ensureFieldAccessorsInitialized( 2887 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder.class); 2888 } 2889 2890 public static com.google.protobuf.Parser<FileUnderConstructionFeature> PARSER = 2891 new com.google.protobuf.AbstractParser<FileUnderConstructionFeature>() { 2892 public FileUnderConstructionFeature parsePartialFrom( 2893 com.google.protobuf.CodedInputStream input, 2894 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2895 throws com.google.protobuf.InvalidProtocolBufferException { 2896 return new FileUnderConstructionFeature(input, extensionRegistry); 2897 } 2898 }; 2899 2900 @java.lang.Override 2901 public com.google.protobuf.Parser<FileUnderConstructionFeature> getParserForType() { 2902 return PARSER; 2903 } 2904 2905 private int bitField0_; 2906 // optional string clientName = 1; 2907 public static final int CLIENTNAME_FIELD_NUMBER = 1; 2908 private java.lang.Object clientName_; 2909 /** 2910 * <code>optional string clientName = 1;</code> 2911 */ 2912 public boolean hasClientName() { 2913 return ((bitField0_ & 0x00000001) == 0x00000001); 2914 } 2915 /** 2916 * <code>optional string clientName = 1;</code> 2917 */ 2918 public java.lang.String getClientName() { 2919 java.lang.Object ref = clientName_; 2920 if (ref instanceof java.lang.String) { 2921 return (java.lang.String) ref; 2922 } else { 2923 com.google.protobuf.ByteString bs = 2924 (com.google.protobuf.ByteString) ref; 2925 java.lang.String s = bs.toStringUtf8(); 2926 if (bs.isValidUtf8()) { 2927 clientName_ = s; 2928 } 2929 return s; 2930 } 2931 } 2932 /** 2933 * <code>optional string clientName = 1;</code> 2934 */ 2935 public com.google.protobuf.ByteString 2936 getClientNameBytes() { 2937 java.lang.Object ref = clientName_; 2938 if (ref instanceof java.lang.String) { 2939 com.google.protobuf.ByteString b = 2940 com.google.protobuf.ByteString.copyFromUtf8( 2941 (java.lang.String) ref); 2942 clientName_ = b; 2943 return b; 2944 } else { 2945 return (com.google.protobuf.ByteString) ref; 2946 } 2947 } 2948 2949 // optional string clientMachine = 2; 2950 public static final int CLIENTMACHINE_FIELD_NUMBER = 2; 2951 private java.lang.Object clientMachine_; 2952 /** 2953 * <code>optional string clientMachine = 2;</code> 2954 */ 2955 public boolean hasClientMachine() { 2956 return ((bitField0_ & 0x00000002) == 0x00000002); 2957 } 2958 /** 2959 * <code>optional string clientMachine = 2;</code> 2960 */ 2961 public java.lang.String getClientMachine() { 2962 java.lang.Object ref = clientMachine_; 2963 if (ref instanceof java.lang.String) { 2964 return (java.lang.String) ref; 2965 } else { 2966 com.google.protobuf.ByteString bs = 2967 (com.google.protobuf.ByteString) ref; 2968 java.lang.String s = bs.toStringUtf8(); 2969 if (bs.isValidUtf8()) { 2970 clientMachine_ = s; 2971 } 2972 return s; 2973 } 2974 } 2975 /** 2976 * <code>optional string clientMachine = 2;</code> 2977 */ 2978 public com.google.protobuf.ByteString 2979 getClientMachineBytes() { 2980 java.lang.Object ref = clientMachine_; 2981 if (ref instanceof java.lang.String) { 2982 com.google.protobuf.ByteString b = 2983 com.google.protobuf.ByteString.copyFromUtf8( 2984 (java.lang.String) ref); 2985 clientMachine_ = b; 2986 return b; 2987 } else { 2988 return (com.google.protobuf.ByteString) ref; 2989 } 2990 } 2991 2992 private void initFields() { 2993 clientName_ = ""; 2994 clientMachine_ = ""; 2995 } 2996 private byte memoizedIsInitialized = -1; 2997 public final boolean isInitialized() { 2998 byte isInitialized = memoizedIsInitialized; 2999 if (isInitialized != -1) return isInitialized == 1; 3000 3001 memoizedIsInitialized = 1; 3002 return true; 3003 } 3004 3005 public void writeTo(com.google.protobuf.CodedOutputStream output) 3006 throws java.io.IOException { 3007 getSerializedSize(); 3008 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3009 output.writeBytes(1, getClientNameBytes()); 3010 } 3011 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3012 output.writeBytes(2, getClientMachineBytes()); 3013 } 3014 getUnknownFields().writeTo(output); 3015 } 3016 3017 private int memoizedSerializedSize = -1; 3018 public int getSerializedSize() { 3019 int size = memoizedSerializedSize; 3020 if (size != -1) return size; 3021 3022 size = 0; 3023 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3024 size += com.google.protobuf.CodedOutputStream 3025 .computeBytesSize(1, getClientNameBytes()); 3026 } 3027 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3028 size += com.google.protobuf.CodedOutputStream 3029 .computeBytesSize(2, getClientMachineBytes()); 3030 } 3031 size += getUnknownFields().getSerializedSize(); 3032 memoizedSerializedSize = size; 3033 return size; 3034 } 3035 3036 private static final long serialVersionUID = 0L; 3037 @java.lang.Override 3038 protected java.lang.Object writeReplace() 3039 throws java.io.ObjectStreamException { 3040 return super.writeReplace(); 3041 } 3042 3043 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3044 com.google.protobuf.ByteString data) 3045 throws com.google.protobuf.InvalidProtocolBufferException { 3046 return PARSER.parseFrom(data); 3047 } 3048 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3049 com.google.protobuf.ByteString data, 3050 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3051 throws com.google.protobuf.InvalidProtocolBufferException { 3052 return PARSER.parseFrom(data, extensionRegistry); 3053 } 3054 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom(byte[] data) 3055 throws com.google.protobuf.InvalidProtocolBufferException { 3056 return PARSER.parseFrom(data); 3057 } 3058 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3059 byte[] data, 3060 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3061 throws com.google.protobuf.InvalidProtocolBufferException { 3062 return PARSER.parseFrom(data, extensionRegistry); 3063 } 3064 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom(java.io.InputStream input) 3065 throws java.io.IOException { 3066 return PARSER.parseFrom(input); 3067 } 3068 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3069 java.io.InputStream input, 3070 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3071 throws java.io.IOException { 3072 return PARSER.parseFrom(input, extensionRegistry); 3073 } 3074 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseDelimitedFrom(java.io.InputStream input) 3075 throws java.io.IOException { 3076 return PARSER.parseDelimitedFrom(input); 3077 } 3078 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseDelimitedFrom( 3079 java.io.InputStream input, 3080 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3081 throws java.io.IOException { 3082 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3083 } 3084 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3085 com.google.protobuf.CodedInputStream input) 3086 throws java.io.IOException { 3087 return PARSER.parseFrom(input); 3088 } 3089 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parseFrom( 3090 com.google.protobuf.CodedInputStream input, 3091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3092 throws java.io.IOException { 3093 return PARSER.parseFrom(input, extensionRegistry); 3094 } 3095 3096 public static Builder newBuilder() { return Builder.create(); } 3097 public Builder newBuilderForType() { return newBuilder(); } 3098 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature prototype) { 3099 return newBuilder().mergeFrom(prototype); 3100 } 3101 public Builder toBuilder() { return newBuilder(this); } 3102 3103 @java.lang.Override 3104 protected Builder newBuilderForType( 3105 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3106 Builder builder = new Builder(parent); 3107 return builder; 3108 } 3109 /** 3110 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature} 3111 * 3112 * <pre> 3113 ** 3114 * under-construction feature for INodeFile 3115 * </pre> 3116 */ 3117 public static final class Builder extends 3118 com.google.protobuf.GeneratedMessage.Builder<Builder> 3119 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder { 3120 public static final com.google.protobuf.Descriptors.Descriptor 3121 getDescriptor() { 3122 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 3123 } 3124 3125 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3126 internalGetFieldAccessorTable() { 3127 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable 3128 .ensureFieldAccessorsInitialized( 3129 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder.class); 3130 } 3131 3132 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.newBuilder() 3133 private Builder() { 3134 maybeForceBuilderInitialization(); 3135 } 3136 3137 private Builder( 3138 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3139 super(parent); 3140 maybeForceBuilderInitialization(); 3141 } 3142 private void maybeForceBuilderInitialization() { 3143 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3144 } 3145 } 3146 private static Builder create() { 3147 return new Builder(); 3148 } 3149 3150 public Builder clear() { 3151 super.clear(); 3152 clientName_ = ""; 3153 bitField0_ = (bitField0_ & ~0x00000001); 3154 clientMachine_ = ""; 3155 bitField0_ = (bitField0_ & ~0x00000002); 3156 return this; 3157 } 3158 3159 public Builder clone() { 3160 return create().mergeFrom(buildPartial()); 3161 } 3162 3163 public com.google.protobuf.Descriptors.Descriptor 3164 getDescriptorForType() { 3165 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 3166 } 3167 3168 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getDefaultInstanceForType() { 3169 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 3170 } 3171 3172 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature build() { 3173 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature result = buildPartial(); 3174 if (!result.isInitialized()) { 3175 throw newUninitializedMessageException(result); 3176 } 3177 return result; 3178 } 3179 3180 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature buildPartial() { 3181 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature(this); 3182 int from_bitField0_ = bitField0_; 3183 int to_bitField0_ = 0; 3184 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 3185 to_bitField0_ |= 0x00000001; 3186 } 3187 result.clientName_ = clientName_; 3188 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 3189 to_bitField0_ |= 0x00000002; 3190 } 3191 result.clientMachine_ = clientMachine_; 3192 result.bitField0_ = to_bitField0_; 3193 onBuilt(); 3194 return result; 3195 } 3196 3197 public Builder mergeFrom(com.google.protobuf.Message other) { 3198 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature) { 3199 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature)other); 3200 } else { 3201 super.mergeFrom(other); 3202 return this; 3203 } 3204 } 3205 3206 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature other) { 3207 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance()) return this; 3208 if (other.hasClientName()) { 3209 bitField0_ |= 0x00000001; 3210 clientName_ = other.clientName_; 3211 onChanged(); 3212 } 3213 if (other.hasClientMachine()) { 3214 bitField0_ |= 0x00000002; 3215 clientMachine_ = other.clientMachine_; 3216 onChanged(); 3217 } 3218 this.mergeUnknownFields(other.getUnknownFields()); 3219 return this; 3220 } 3221 3222 public final boolean isInitialized() { 3223 return true; 3224 } 3225 3226 public Builder mergeFrom( 3227 com.google.protobuf.CodedInputStream input, 3228 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3229 throws java.io.IOException { 3230 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature parsedMessage = null; 3231 try { 3232 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3233 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3234 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature) e.getUnfinishedMessage(); 3235 throw e; 3236 } finally { 3237 if (parsedMessage != null) { 3238 mergeFrom(parsedMessage); 3239 } 3240 } 3241 return this; 3242 } 3243 private int bitField0_; 3244 3245 // optional string clientName = 1; 3246 private java.lang.Object clientName_ = ""; 3247 /** 3248 * <code>optional string clientName = 1;</code> 3249 */ 3250 public boolean hasClientName() { 3251 return ((bitField0_ & 0x00000001) == 0x00000001); 3252 } 3253 /** 3254 * <code>optional string clientName = 1;</code> 3255 */ 3256 public java.lang.String getClientName() { 3257 java.lang.Object ref = clientName_; 3258 if (!(ref instanceof java.lang.String)) { 3259 java.lang.String s = ((com.google.protobuf.ByteString) ref) 3260 .toStringUtf8(); 3261 clientName_ = s; 3262 return s; 3263 } else { 3264 return (java.lang.String) ref; 3265 } 3266 } 3267 /** 3268 * <code>optional string clientName = 1;</code> 3269 */ 3270 public com.google.protobuf.ByteString 3271 getClientNameBytes() { 3272 java.lang.Object ref = clientName_; 3273 if (ref instanceof String) { 3274 com.google.protobuf.ByteString b = 3275 com.google.protobuf.ByteString.copyFromUtf8( 3276 (java.lang.String) ref); 3277 clientName_ = b; 3278 return b; 3279 } else { 3280 return (com.google.protobuf.ByteString) ref; 3281 } 3282 } 3283 /** 3284 * <code>optional string clientName = 1;</code> 3285 */ 3286 public Builder setClientName( 3287 java.lang.String value) { 3288 if (value == null) { 3289 throw new NullPointerException(); 3290 } 3291 bitField0_ |= 0x00000001; 3292 clientName_ = value; 3293 onChanged(); 3294 return this; 3295 } 3296 /** 3297 * <code>optional string clientName = 1;</code> 3298 */ 3299 public Builder clearClientName() { 3300 bitField0_ = (bitField0_ & ~0x00000001); 3301 clientName_ = getDefaultInstance().getClientName(); 3302 onChanged(); 3303 return this; 3304 } 3305 /** 3306 * <code>optional string clientName = 1;</code> 3307 */ 3308 public Builder setClientNameBytes( 3309 com.google.protobuf.ByteString value) { 3310 if (value == null) { 3311 throw new NullPointerException(); 3312 } 3313 bitField0_ |= 0x00000001; 3314 clientName_ = value; 3315 onChanged(); 3316 return this; 3317 } 3318 3319 // optional string clientMachine = 2; 3320 private java.lang.Object clientMachine_ = ""; 3321 /** 3322 * <code>optional string clientMachine = 2;</code> 3323 */ 3324 public boolean hasClientMachine() { 3325 return ((bitField0_ & 0x00000002) == 0x00000002); 3326 } 3327 /** 3328 * <code>optional string clientMachine = 2;</code> 3329 */ 3330 public java.lang.String getClientMachine() { 3331 java.lang.Object ref = clientMachine_; 3332 if (!(ref instanceof java.lang.String)) { 3333 java.lang.String s = ((com.google.protobuf.ByteString) ref) 3334 .toStringUtf8(); 3335 clientMachine_ = s; 3336 return s; 3337 } else { 3338 return (java.lang.String) ref; 3339 } 3340 } 3341 /** 3342 * <code>optional string clientMachine = 2;</code> 3343 */ 3344 public com.google.protobuf.ByteString 3345 getClientMachineBytes() { 3346 java.lang.Object ref = clientMachine_; 3347 if (ref instanceof String) { 3348 com.google.protobuf.ByteString b = 3349 com.google.protobuf.ByteString.copyFromUtf8( 3350 (java.lang.String) ref); 3351 clientMachine_ = b; 3352 return b; 3353 } else { 3354 return (com.google.protobuf.ByteString) ref; 3355 } 3356 } 3357 /** 3358 * <code>optional string clientMachine = 2;</code> 3359 */ 3360 public Builder setClientMachine( 3361 java.lang.String value) { 3362 if (value == null) { 3363 throw new NullPointerException(); 3364 } 3365 bitField0_ |= 0x00000002; 3366 clientMachine_ = value; 3367 onChanged(); 3368 return this; 3369 } 3370 /** 3371 * <code>optional string clientMachine = 2;</code> 3372 */ 3373 public Builder clearClientMachine() { 3374 bitField0_ = (bitField0_ & ~0x00000002); 3375 clientMachine_ = getDefaultInstance().getClientMachine(); 3376 onChanged(); 3377 return this; 3378 } 3379 /** 3380 * <code>optional string clientMachine = 2;</code> 3381 */ 3382 public Builder setClientMachineBytes( 3383 com.google.protobuf.ByteString value) { 3384 if (value == null) { 3385 throw new NullPointerException(); 3386 } 3387 bitField0_ |= 0x00000002; 3388 clientMachine_ = value; 3389 onChanged(); 3390 return this; 3391 } 3392 3393 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature) 3394 } 3395 3396 static { 3397 defaultInstance = new FileUnderConstructionFeature(true); 3398 defaultInstance.initFields(); 3399 } 3400 3401 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature) 3402 } 3403 3404 public interface AclFeatureProtoOrBuilder 3405 extends com.google.protobuf.MessageOrBuilder { 3406 3407 // repeated fixed32 entries = 2 [packed = true]; 3408 /** 3409 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3410 * 3411 * <pre> 3412 ** 3413 * An ACL entry is represented by a 32-bit integer in Big Endian 3414 * format. The bits can be divided in four segments: 3415 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3416 * 3417 * [0:2) -- reserved for futute uses. 3418 * [2:26) -- the name of the entry, which is an ID that points to a 3419 * string in the StringTableSection. 3420 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3421 * [27:29) -- the type of the entry (AclEntryTypeProto) 3422 * [29:32) -- the permission of the entry (FsActionProto) 3423 * </pre> 3424 */ 3425 java.util.List<java.lang.Integer> getEntriesList(); 3426 /** 3427 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3428 * 3429 * <pre> 3430 ** 3431 * An ACL entry is represented by a 32-bit integer in Big Endian 3432 * format. The bits can be divided in four segments: 3433 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3434 * 3435 * [0:2) -- reserved for futute uses. 3436 * [2:26) -- the name of the entry, which is an ID that points to a 3437 * string in the StringTableSection. 3438 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3439 * [27:29) -- the type of the entry (AclEntryTypeProto) 3440 * [29:32) -- the permission of the entry (FsActionProto) 3441 * </pre> 3442 */ 3443 int getEntriesCount(); 3444 /** 3445 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3446 * 3447 * <pre> 3448 ** 3449 * An ACL entry is represented by a 32-bit integer in Big Endian 3450 * format. The bits can be divided in four segments: 3451 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3452 * 3453 * [0:2) -- reserved for futute uses. 3454 * [2:26) -- the name of the entry, which is an ID that points to a 3455 * string in the StringTableSection. 3456 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3457 * [27:29) -- the type of the entry (AclEntryTypeProto) 3458 * [29:32) -- the permission of the entry (FsActionProto) 3459 * </pre> 3460 */ 3461 int getEntries(int index); 3462 } 3463 /** 3464 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.AclFeatureProto} 3465 */ 3466 public static final class AclFeatureProto extends 3467 com.google.protobuf.GeneratedMessage 3468 implements AclFeatureProtoOrBuilder { 3469 // Use AclFeatureProto.newBuilder() to construct. 3470 private AclFeatureProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3471 super(builder); 3472 this.unknownFields = builder.getUnknownFields(); 3473 } 3474 private AclFeatureProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3475 3476 private static final AclFeatureProto defaultInstance; 3477 public static AclFeatureProto getDefaultInstance() { 3478 return defaultInstance; 3479 } 3480 3481 public AclFeatureProto getDefaultInstanceForType() { 3482 return defaultInstance; 3483 } 3484 3485 private final com.google.protobuf.UnknownFieldSet unknownFields; 3486 @java.lang.Override 3487 public final com.google.protobuf.UnknownFieldSet 3488 getUnknownFields() { 3489 return this.unknownFields; 3490 } 3491 private AclFeatureProto( 3492 com.google.protobuf.CodedInputStream input, 3493 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3494 throws com.google.protobuf.InvalidProtocolBufferException { 3495 initFields(); 3496 int mutable_bitField0_ = 0; 3497 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3498 com.google.protobuf.UnknownFieldSet.newBuilder(); 3499 try { 3500 boolean done = false; 3501 while (!done) { 3502 int tag = input.readTag(); 3503 switch (tag) { 3504 case 0: 3505 done = true; 3506 break; 3507 default: { 3508 if (!parseUnknownField(input, unknownFields, 3509 extensionRegistry, tag)) { 3510 done = true; 3511 } 3512 break; 3513 } 3514 case 21: { 3515 if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 3516 entries_ = new java.util.ArrayList<java.lang.Integer>(); 3517 mutable_bitField0_ |= 0x00000001; 3518 } 3519 entries_.add(input.readFixed32()); 3520 break; 3521 } 3522 case 18: { 3523 int length = input.readRawVarint32(); 3524 int limit = input.pushLimit(length); 3525 if (!((mutable_bitField0_ & 0x00000001) == 0x00000001) && input.getBytesUntilLimit() > 0) { 3526 entries_ = new java.util.ArrayList<java.lang.Integer>(); 3527 mutable_bitField0_ |= 0x00000001; 3528 } 3529 while (input.getBytesUntilLimit() > 0) { 3530 entries_.add(input.readFixed32()); 3531 } 3532 input.popLimit(limit); 3533 break; 3534 } 3535 } 3536 } 3537 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3538 throw e.setUnfinishedMessage(this); 3539 } catch (java.io.IOException e) { 3540 throw new com.google.protobuf.InvalidProtocolBufferException( 3541 e.getMessage()).setUnfinishedMessage(this); 3542 } finally { 3543 if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 3544 entries_ = java.util.Collections.unmodifiableList(entries_); 3545 } 3546 this.unknownFields = unknownFields.build(); 3547 makeExtensionsImmutable(); 3548 } 3549 } 3550 public static final com.google.protobuf.Descriptors.Descriptor 3551 getDescriptor() { 3552 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3553 } 3554 3555 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3556 internalGetFieldAccessorTable() { 3557 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable 3558 .ensureFieldAccessorsInitialized( 3559 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder.class); 3560 } 3561 3562 public static com.google.protobuf.Parser<AclFeatureProto> PARSER = 3563 new com.google.protobuf.AbstractParser<AclFeatureProto>() { 3564 public AclFeatureProto parsePartialFrom( 3565 com.google.protobuf.CodedInputStream input, 3566 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3567 throws com.google.protobuf.InvalidProtocolBufferException { 3568 return new AclFeatureProto(input, extensionRegistry); 3569 } 3570 }; 3571 3572 @java.lang.Override 3573 public com.google.protobuf.Parser<AclFeatureProto> getParserForType() { 3574 return PARSER; 3575 } 3576 3577 // repeated fixed32 entries = 2 [packed = true]; 3578 public static final int ENTRIES_FIELD_NUMBER = 2; 3579 private java.util.List<java.lang.Integer> entries_; 3580 /** 3581 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3582 * 3583 * <pre> 3584 ** 3585 * An ACL entry is represented by a 32-bit integer in Big Endian 3586 * format. The bits can be divided in four segments: 3587 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3588 * 3589 * [0:2) -- reserved for futute uses. 3590 * [2:26) -- the name of the entry, which is an ID that points to a 3591 * string in the StringTableSection. 3592 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3593 * [27:29) -- the type of the entry (AclEntryTypeProto) 3594 * [29:32) -- the permission of the entry (FsActionProto) 3595 * </pre> 3596 */ 3597 public java.util.List<java.lang.Integer> 3598 getEntriesList() { 3599 return entries_; 3600 } 3601 /** 3602 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3603 * 3604 * <pre> 3605 ** 3606 * An ACL entry is represented by a 32-bit integer in Big Endian 3607 * format. The bits can be divided in four segments: 3608 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3609 * 3610 * [0:2) -- reserved for futute uses. 3611 * [2:26) -- the name of the entry, which is an ID that points to a 3612 * string in the StringTableSection. 3613 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3614 * [27:29) -- the type of the entry (AclEntryTypeProto) 3615 * [29:32) -- the permission of the entry (FsActionProto) 3616 * </pre> 3617 */ 3618 public int getEntriesCount() { 3619 return entries_.size(); 3620 } 3621 /** 3622 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3623 * 3624 * <pre> 3625 ** 3626 * An ACL entry is represented by a 32-bit integer in Big Endian 3627 * format. The bits can be divided in four segments: 3628 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3629 * 3630 * [0:2) -- reserved for futute uses. 3631 * [2:26) -- the name of the entry, which is an ID that points to a 3632 * string in the StringTableSection. 3633 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3634 * [27:29) -- the type of the entry (AclEntryTypeProto) 3635 * [29:32) -- the permission of the entry (FsActionProto) 3636 * </pre> 3637 */ 3638 public int getEntries(int index) { 3639 return entries_.get(index); 3640 } 3641 private int entriesMemoizedSerializedSize = -1; 3642 3643 private void initFields() { 3644 entries_ = java.util.Collections.emptyList(); 3645 } 3646 private byte memoizedIsInitialized = -1; 3647 public final boolean isInitialized() { 3648 byte isInitialized = memoizedIsInitialized; 3649 if (isInitialized != -1) return isInitialized == 1; 3650 3651 memoizedIsInitialized = 1; 3652 return true; 3653 } 3654 3655 public void writeTo(com.google.protobuf.CodedOutputStream output) 3656 throws java.io.IOException { 3657 getSerializedSize(); 3658 if (getEntriesList().size() > 0) { 3659 output.writeRawVarint32(18); 3660 output.writeRawVarint32(entriesMemoizedSerializedSize); 3661 } 3662 for (int i = 0; i < entries_.size(); i++) { 3663 output.writeFixed32NoTag(entries_.get(i)); 3664 } 3665 getUnknownFields().writeTo(output); 3666 } 3667 3668 private int memoizedSerializedSize = -1; 3669 public int getSerializedSize() { 3670 int size = memoizedSerializedSize; 3671 if (size != -1) return size; 3672 3673 size = 0; 3674 { 3675 int dataSize = 0; 3676 dataSize = 4 * getEntriesList().size(); 3677 size += dataSize; 3678 if (!getEntriesList().isEmpty()) { 3679 size += 1; 3680 size += com.google.protobuf.CodedOutputStream 3681 .computeInt32SizeNoTag(dataSize); 3682 } 3683 entriesMemoizedSerializedSize = dataSize; 3684 } 3685 size += getUnknownFields().getSerializedSize(); 3686 memoizedSerializedSize = size; 3687 return size; 3688 } 3689 3690 private static final long serialVersionUID = 0L; 3691 @java.lang.Override 3692 protected java.lang.Object writeReplace() 3693 throws java.io.ObjectStreamException { 3694 return super.writeReplace(); 3695 } 3696 3697 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3698 com.google.protobuf.ByteString data) 3699 throws com.google.protobuf.InvalidProtocolBufferException { 3700 return PARSER.parseFrom(data); 3701 } 3702 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3703 com.google.protobuf.ByteString data, 3704 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3705 throws com.google.protobuf.InvalidProtocolBufferException { 3706 return PARSER.parseFrom(data, extensionRegistry); 3707 } 3708 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom(byte[] data) 3709 throws com.google.protobuf.InvalidProtocolBufferException { 3710 return PARSER.parseFrom(data); 3711 } 3712 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3713 byte[] data, 3714 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3715 throws com.google.protobuf.InvalidProtocolBufferException { 3716 return PARSER.parseFrom(data, extensionRegistry); 3717 } 3718 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom(java.io.InputStream input) 3719 throws java.io.IOException { 3720 return PARSER.parseFrom(input); 3721 } 3722 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3723 java.io.InputStream input, 3724 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3725 throws java.io.IOException { 3726 return PARSER.parseFrom(input, extensionRegistry); 3727 } 3728 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseDelimitedFrom(java.io.InputStream input) 3729 throws java.io.IOException { 3730 return PARSER.parseDelimitedFrom(input); 3731 } 3732 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseDelimitedFrom( 3733 java.io.InputStream input, 3734 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3735 throws java.io.IOException { 3736 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3737 } 3738 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3739 com.google.protobuf.CodedInputStream input) 3740 throws java.io.IOException { 3741 return PARSER.parseFrom(input); 3742 } 3743 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parseFrom( 3744 com.google.protobuf.CodedInputStream input, 3745 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3746 throws java.io.IOException { 3747 return PARSER.parseFrom(input, extensionRegistry); 3748 } 3749 3750 public static Builder newBuilder() { return Builder.create(); } 3751 public Builder newBuilderForType() { return newBuilder(); } 3752 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto prototype) { 3753 return newBuilder().mergeFrom(prototype); 3754 } 3755 public Builder toBuilder() { return newBuilder(this); } 3756 3757 @java.lang.Override 3758 protected Builder newBuilderForType( 3759 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3760 Builder builder = new Builder(parent); 3761 return builder; 3762 } 3763 /** 3764 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.AclFeatureProto} 3765 */ 3766 public static final class Builder extends 3767 com.google.protobuf.GeneratedMessage.Builder<Builder> 3768 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder { 3769 public static final com.google.protobuf.Descriptors.Descriptor 3770 getDescriptor() { 3771 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3772 } 3773 3774 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3775 internalGetFieldAccessorTable() { 3776 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable 3777 .ensureFieldAccessorsInitialized( 3778 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder.class); 3779 } 3780 3781 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder() 3782 private Builder() { 3783 maybeForceBuilderInitialization(); 3784 } 3785 3786 private Builder( 3787 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3788 super(parent); 3789 maybeForceBuilderInitialization(); 3790 } 3791 private void maybeForceBuilderInitialization() { 3792 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3793 } 3794 } 3795 private static Builder create() { 3796 return new Builder(); 3797 } 3798 3799 public Builder clear() { 3800 super.clear(); 3801 entries_ = java.util.Collections.emptyList(); 3802 bitField0_ = (bitField0_ & ~0x00000001); 3803 return this; 3804 } 3805 3806 public Builder clone() { 3807 return create().mergeFrom(buildPartial()); 3808 } 3809 3810 public com.google.protobuf.Descriptors.Descriptor 3811 getDescriptorForType() { 3812 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 3813 } 3814 3815 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getDefaultInstanceForType() { 3816 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 3817 } 3818 3819 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto build() { 3820 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto result = buildPartial(); 3821 if (!result.isInitialized()) { 3822 throw newUninitializedMessageException(result); 3823 } 3824 return result; 3825 } 3826 3827 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto buildPartial() { 3828 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto(this); 3829 int from_bitField0_ = bitField0_; 3830 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3831 entries_ = java.util.Collections.unmodifiableList(entries_); 3832 bitField0_ = (bitField0_ & ~0x00000001); 3833 } 3834 result.entries_ = entries_; 3835 onBuilt(); 3836 return result; 3837 } 3838 3839 public Builder mergeFrom(com.google.protobuf.Message other) { 3840 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto) { 3841 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto)other); 3842 } else { 3843 super.mergeFrom(other); 3844 return this; 3845 } 3846 } 3847 3848 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto other) { 3849 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) return this; 3850 if (!other.entries_.isEmpty()) { 3851 if (entries_.isEmpty()) { 3852 entries_ = other.entries_; 3853 bitField0_ = (bitField0_ & ~0x00000001); 3854 } else { 3855 ensureEntriesIsMutable(); 3856 entries_.addAll(other.entries_); 3857 } 3858 onChanged(); 3859 } 3860 this.mergeUnknownFields(other.getUnknownFields()); 3861 return this; 3862 } 3863 3864 public final boolean isInitialized() { 3865 return true; 3866 } 3867 3868 public Builder mergeFrom( 3869 com.google.protobuf.CodedInputStream input, 3870 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3871 throws java.io.IOException { 3872 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto parsedMessage = null; 3873 try { 3874 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3875 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3876 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto) e.getUnfinishedMessage(); 3877 throw e; 3878 } finally { 3879 if (parsedMessage != null) { 3880 mergeFrom(parsedMessage); 3881 } 3882 } 3883 return this; 3884 } 3885 private int bitField0_; 3886 3887 // repeated fixed32 entries = 2 [packed = true]; 3888 private java.util.List<java.lang.Integer> entries_ = java.util.Collections.emptyList(); 3889 private void ensureEntriesIsMutable() { 3890 if (!((bitField0_ & 0x00000001) == 0x00000001)) { 3891 entries_ = new java.util.ArrayList<java.lang.Integer>(entries_); 3892 bitField0_ |= 0x00000001; 3893 } 3894 } 3895 /** 3896 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3897 * 3898 * <pre> 3899 ** 3900 * An ACL entry is represented by a 32-bit integer in Big Endian 3901 * format. The bits can be divided in four segments: 3902 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3903 * 3904 * [0:2) -- reserved for futute uses. 3905 * [2:26) -- the name of the entry, which is an ID that points to a 3906 * string in the StringTableSection. 3907 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3908 * [27:29) -- the type of the entry (AclEntryTypeProto) 3909 * [29:32) -- the permission of the entry (FsActionProto) 3910 * </pre> 3911 */ 3912 public java.util.List<java.lang.Integer> 3913 getEntriesList() { 3914 return java.util.Collections.unmodifiableList(entries_); 3915 } 3916 /** 3917 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3918 * 3919 * <pre> 3920 ** 3921 * An ACL entry is represented by a 32-bit integer in Big Endian 3922 * format. The bits can be divided in four segments: 3923 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3924 * 3925 * [0:2) -- reserved for futute uses. 3926 * [2:26) -- the name of the entry, which is an ID that points to a 3927 * string in the StringTableSection. 3928 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3929 * [27:29) -- the type of the entry (AclEntryTypeProto) 3930 * [29:32) -- the permission of the entry (FsActionProto) 3931 * </pre> 3932 */ 3933 public int getEntriesCount() { 3934 return entries_.size(); 3935 } 3936 /** 3937 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3938 * 3939 * <pre> 3940 ** 3941 * An ACL entry is represented by a 32-bit integer in Big Endian 3942 * format. The bits can be divided in four segments: 3943 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3944 * 3945 * [0:2) -- reserved for futute uses. 3946 * [2:26) -- the name of the entry, which is an ID that points to a 3947 * string in the StringTableSection. 3948 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3949 * [27:29) -- the type of the entry (AclEntryTypeProto) 3950 * [29:32) -- the permission of the entry (FsActionProto) 3951 * </pre> 3952 */ 3953 public int getEntries(int index) { 3954 return entries_.get(index); 3955 } 3956 /** 3957 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3958 * 3959 * <pre> 3960 ** 3961 * An ACL entry is represented by a 32-bit integer in Big Endian 3962 * format. The bits can be divided in four segments: 3963 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3964 * 3965 * [0:2) -- reserved for futute uses. 3966 * [2:26) -- the name of the entry, which is an ID that points to a 3967 * string in the StringTableSection. 3968 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3969 * [27:29) -- the type of the entry (AclEntryTypeProto) 3970 * [29:32) -- the permission of the entry (FsActionProto) 3971 * </pre> 3972 */ 3973 public Builder setEntries( 3974 int index, int value) { 3975 ensureEntriesIsMutable(); 3976 entries_.set(index, value); 3977 onChanged(); 3978 return this; 3979 } 3980 /** 3981 * <code>repeated fixed32 entries = 2 [packed = true];</code> 3982 * 3983 * <pre> 3984 ** 3985 * An ACL entry is represented by a 32-bit integer in Big Endian 3986 * format. The bits can be divided in four segments: 3987 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 3988 * 3989 * [0:2) -- reserved for futute uses. 3990 * [2:26) -- the name of the entry, which is an ID that points to a 3991 * string in the StringTableSection. 3992 * [26:27) -- the scope of the entry (AclEntryScopeProto) 3993 * [27:29) -- the type of the entry (AclEntryTypeProto) 3994 * [29:32) -- the permission of the entry (FsActionProto) 3995 * </pre> 3996 */ 3997 public Builder addEntries(int value) { 3998 ensureEntriesIsMutable(); 3999 entries_.add(value); 4000 onChanged(); 4001 return this; 4002 } 4003 /** 4004 * <code>repeated fixed32 entries = 2 [packed = true];</code> 4005 * 4006 * <pre> 4007 ** 4008 * An ACL entry is represented by a 32-bit integer in Big Endian 4009 * format. The bits can be divided in four segments: 4010 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 4011 * 4012 * [0:2) -- reserved for futute uses. 4013 * [2:26) -- the name of the entry, which is an ID that points to a 4014 * string in the StringTableSection. 4015 * [26:27) -- the scope of the entry (AclEntryScopeProto) 4016 * [27:29) -- the type of the entry (AclEntryTypeProto) 4017 * [29:32) -- the permission of the entry (FsActionProto) 4018 * </pre> 4019 */ 4020 public Builder addAllEntries( 4021 java.lang.Iterable<? extends java.lang.Integer> values) { 4022 ensureEntriesIsMutable(); 4023 super.addAll(values, entries_); 4024 onChanged(); 4025 return this; 4026 } 4027 /** 4028 * <code>repeated fixed32 entries = 2 [packed = true];</code> 4029 * 4030 * <pre> 4031 ** 4032 * An ACL entry is represented by a 32-bit integer in Big Endian 4033 * format. The bits can be divided in four segments: 4034 * [0:2) || [2:26) || [26:27) || [27:29) || [29:32) 4035 * 4036 * [0:2) -- reserved for futute uses. 4037 * [2:26) -- the name of the entry, which is an ID that points to a 4038 * string in the StringTableSection. 4039 * [26:27) -- the scope of the entry (AclEntryScopeProto) 4040 * [27:29) -- the type of the entry (AclEntryTypeProto) 4041 * [29:32) -- the permission of the entry (FsActionProto) 4042 * </pre> 4043 */ 4044 public Builder clearEntries() { 4045 entries_ = java.util.Collections.emptyList(); 4046 bitField0_ = (bitField0_ & ~0x00000001); 4047 onChanged(); 4048 return this; 4049 } 4050 4051 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.AclFeatureProto) 4052 } 4053 4054 static { 4055 defaultInstance = new AclFeatureProto(true); 4056 defaultInstance.initFields(); 4057 } 4058 4059 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.AclFeatureProto) 4060 } 4061 4062 public interface XAttrCompactProtoOrBuilder 4063 extends com.google.protobuf.MessageOrBuilder { 4064 4065 // required fixed32 name = 1; 4066 /** 4067 * <code>required fixed32 name = 1;</code> 4068 * 4069 * <pre> 4070 ** 4071 * 4072 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4073 * [2:26) -- the name of the entry, which is an ID that points to a 4074 * string in the StringTableSection. 4075 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4076 * so only 2 bits were needed. At that time, this bit was reserved. When a 4077 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4078 * bit. 4079 * [27:32) -- reserved for future uses. 4080 * </pre> 4081 */ 4082 boolean hasName(); 4083 /** 4084 * <code>required fixed32 name = 1;</code> 4085 * 4086 * <pre> 4087 ** 4088 * 4089 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4090 * [2:26) -- the name of the entry, which is an ID that points to a 4091 * string in the StringTableSection. 4092 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4093 * so only 2 bits were needed. At that time, this bit was reserved. When a 4094 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4095 * bit. 4096 * [27:32) -- reserved for future uses. 4097 * </pre> 4098 */ 4099 int getName(); 4100 4101 // optional bytes value = 2; 4102 /** 4103 * <code>optional bytes value = 2;</code> 4104 */ 4105 boolean hasValue(); 4106 /** 4107 * <code>optional bytes value = 2;</code> 4108 */ 4109 com.google.protobuf.ByteString getValue(); 4110 } 4111 /** 4112 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto} 4113 */ 4114 public static final class XAttrCompactProto extends 4115 com.google.protobuf.GeneratedMessage 4116 implements XAttrCompactProtoOrBuilder { 4117 // Use XAttrCompactProto.newBuilder() to construct. 4118 private XAttrCompactProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4119 super(builder); 4120 this.unknownFields = builder.getUnknownFields(); 4121 } 4122 private XAttrCompactProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4123 4124 private static final XAttrCompactProto defaultInstance; 4125 public static XAttrCompactProto getDefaultInstance() { 4126 return defaultInstance; 4127 } 4128 4129 public XAttrCompactProto getDefaultInstanceForType() { 4130 return defaultInstance; 4131 } 4132 4133 private final com.google.protobuf.UnknownFieldSet unknownFields; 4134 @java.lang.Override 4135 public final com.google.protobuf.UnknownFieldSet 4136 getUnknownFields() { 4137 return this.unknownFields; 4138 } 4139 private XAttrCompactProto( 4140 com.google.protobuf.CodedInputStream input, 4141 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4142 throws com.google.protobuf.InvalidProtocolBufferException { 4143 initFields(); 4144 int mutable_bitField0_ = 0; 4145 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4146 com.google.protobuf.UnknownFieldSet.newBuilder(); 4147 try { 4148 boolean done = false; 4149 while (!done) { 4150 int tag = input.readTag(); 4151 switch (tag) { 4152 case 0: 4153 done = true; 4154 break; 4155 default: { 4156 if (!parseUnknownField(input, unknownFields, 4157 extensionRegistry, tag)) { 4158 done = true; 4159 } 4160 break; 4161 } 4162 case 13: { 4163 bitField0_ |= 0x00000001; 4164 name_ = input.readFixed32(); 4165 break; 4166 } 4167 case 18: { 4168 bitField0_ |= 0x00000002; 4169 value_ = input.readBytes(); 4170 break; 4171 } 4172 } 4173 } 4174 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4175 throw e.setUnfinishedMessage(this); 4176 } catch (java.io.IOException e) { 4177 throw new com.google.protobuf.InvalidProtocolBufferException( 4178 e.getMessage()).setUnfinishedMessage(this); 4179 } finally { 4180 this.unknownFields = unknownFields.build(); 4181 makeExtensionsImmutable(); 4182 } 4183 } 4184 public static final com.google.protobuf.Descriptors.Descriptor 4185 getDescriptor() { 4186 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor; 4187 } 4188 4189 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4190 internalGetFieldAccessorTable() { 4191 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_fieldAccessorTable 4192 .ensureFieldAccessorsInitialized( 4193 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder.class); 4194 } 4195 4196 public static com.google.protobuf.Parser<XAttrCompactProto> PARSER = 4197 new com.google.protobuf.AbstractParser<XAttrCompactProto>() { 4198 public XAttrCompactProto parsePartialFrom( 4199 com.google.protobuf.CodedInputStream input, 4200 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4201 throws com.google.protobuf.InvalidProtocolBufferException { 4202 return new XAttrCompactProto(input, extensionRegistry); 4203 } 4204 }; 4205 4206 @java.lang.Override 4207 public com.google.protobuf.Parser<XAttrCompactProto> getParserForType() { 4208 return PARSER; 4209 } 4210 4211 private int bitField0_; 4212 // required fixed32 name = 1; 4213 public static final int NAME_FIELD_NUMBER = 1; 4214 private int name_; 4215 /** 4216 * <code>required fixed32 name = 1;</code> 4217 * 4218 * <pre> 4219 ** 4220 * 4221 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4222 * [2:26) -- the name of the entry, which is an ID that points to a 4223 * string in the StringTableSection. 4224 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4225 * so only 2 bits were needed. At that time, this bit was reserved. When a 4226 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4227 * bit. 4228 * [27:32) -- reserved for future uses. 4229 * </pre> 4230 */ 4231 public boolean hasName() { 4232 return ((bitField0_ & 0x00000001) == 0x00000001); 4233 } 4234 /** 4235 * <code>required fixed32 name = 1;</code> 4236 * 4237 * <pre> 4238 ** 4239 * 4240 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4241 * [2:26) -- the name of the entry, which is an ID that points to a 4242 * string in the StringTableSection. 4243 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4244 * so only 2 bits were needed. At that time, this bit was reserved. When a 4245 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4246 * bit. 4247 * [27:32) -- reserved for future uses. 4248 * </pre> 4249 */ 4250 public int getName() { 4251 return name_; 4252 } 4253 4254 // optional bytes value = 2; 4255 public static final int VALUE_FIELD_NUMBER = 2; 4256 private com.google.protobuf.ByteString value_; 4257 /** 4258 * <code>optional bytes value = 2;</code> 4259 */ 4260 public boolean hasValue() { 4261 return ((bitField0_ & 0x00000002) == 0x00000002); 4262 } 4263 /** 4264 * <code>optional bytes value = 2;</code> 4265 */ 4266 public com.google.protobuf.ByteString getValue() { 4267 return value_; 4268 } 4269 4270 private void initFields() { 4271 name_ = 0; 4272 value_ = com.google.protobuf.ByteString.EMPTY; 4273 } 4274 private byte memoizedIsInitialized = -1; 4275 public final boolean isInitialized() { 4276 byte isInitialized = memoizedIsInitialized; 4277 if (isInitialized != -1) return isInitialized == 1; 4278 4279 if (!hasName()) { 4280 memoizedIsInitialized = 0; 4281 return false; 4282 } 4283 memoizedIsInitialized = 1; 4284 return true; 4285 } 4286 4287 public void writeTo(com.google.protobuf.CodedOutputStream output) 4288 throws java.io.IOException { 4289 getSerializedSize(); 4290 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4291 output.writeFixed32(1, name_); 4292 } 4293 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4294 output.writeBytes(2, value_); 4295 } 4296 getUnknownFields().writeTo(output); 4297 } 4298 4299 private int memoizedSerializedSize = -1; 4300 public int getSerializedSize() { 4301 int size = memoizedSerializedSize; 4302 if (size != -1) return size; 4303 4304 size = 0; 4305 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4306 size += com.google.protobuf.CodedOutputStream 4307 .computeFixed32Size(1, name_); 4308 } 4309 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4310 size += com.google.protobuf.CodedOutputStream 4311 .computeBytesSize(2, value_); 4312 } 4313 size += getUnknownFields().getSerializedSize(); 4314 memoizedSerializedSize = size; 4315 return size; 4316 } 4317 4318 private static final long serialVersionUID = 0L; 4319 @java.lang.Override 4320 protected java.lang.Object writeReplace() 4321 throws java.io.ObjectStreamException { 4322 return super.writeReplace(); 4323 } 4324 4325 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4326 com.google.protobuf.ByteString data) 4327 throws com.google.protobuf.InvalidProtocolBufferException { 4328 return PARSER.parseFrom(data); 4329 } 4330 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4331 com.google.protobuf.ByteString data, 4332 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4333 throws com.google.protobuf.InvalidProtocolBufferException { 4334 return PARSER.parseFrom(data, extensionRegistry); 4335 } 4336 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom(byte[] data) 4337 throws com.google.protobuf.InvalidProtocolBufferException { 4338 return PARSER.parseFrom(data); 4339 } 4340 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4341 byte[] data, 4342 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4343 throws com.google.protobuf.InvalidProtocolBufferException { 4344 return PARSER.parseFrom(data, extensionRegistry); 4345 } 4346 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom(java.io.InputStream input) 4347 throws java.io.IOException { 4348 return PARSER.parseFrom(input); 4349 } 4350 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4351 java.io.InputStream input, 4352 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4353 throws java.io.IOException { 4354 return PARSER.parseFrom(input, extensionRegistry); 4355 } 4356 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseDelimitedFrom(java.io.InputStream input) 4357 throws java.io.IOException { 4358 return PARSER.parseDelimitedFrom(input); 4359 } 4360 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseDelimitedFrom( 4361 java.io.InputStream input, 4362 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4363 throws java.io.IOException { 4364 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4365 } 4366 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4367 com.google.protobuf.CodedInputStream input) 4368 throws java.io.IOException { 4369 return PARSER.parseFrom(input); 4370 } 4371 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parseFrom( 4372 com.google.protobuf.CodedInputStream input, 4373 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4374 throws java.io.IOException { 4375 return PARSER.parseFrom(input, extensionRegistry); 4376 } 4377 4378 public static Builder newBuilder() { return Builder.create(); } 4379 public Builder newBuilderForType() { return newBuilder(); } 4380 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto prototype) { 4381 return newBuilder().mergeFrom(prototype); 4382 } 4383 public Builder toBuilder() { return newBuilder(this); } 4384 4385 @java.lang.Override 4386 protected Builder newBuilderForType( 4387 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4388 Builder builder = new Builder(parent); 4389 return builder; 4390 } 4391 /** 4392 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto} 4393 */ 4394 public static final class Builder extends 4395 com.google.protobuf.GeneratedMessage.Builder<Builder> 4396 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder { 4397 public static final com.google.protobuf.Descriptors.Descriptor 4398 getDescriptor() { 4399 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor; 4400 } 4401 4402 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4403 internalGetFieldAccessorTable() { 4404 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_fieldAccessorTable 4405 .ensureFieldAccessorsInitialized( 4406 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder.class); 4407 } 4408 4409 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.newBuilder() 4410 private Builder() { 4411 maybeForceBuilderInitialization(); 4412 } 4413 4414 private Builder( 4415 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4416 super(parent); 4417 maybeForceBuilderInitialization(); 4418 } 4419 private void maybeForceBuilderInitialization() { 4420 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4421 } 4422 } 4423 private static Builder create() { 4424 return new Builder(); 4425 } 4426 4427 public Builder clear() { 4428 super.clear(); 4429 name_ = 0; 4430 bitField0_ = (bitField0_ & ~0x00000001); 4431 value_ = com.google.protobuf.ByteString.EMPTY; 4432 bitField0_ = (bitField0_ & ~0x00000002); 4433 return this; 4434 } 4435 4436 public Builder clone() { 4437 return create().mergeFrom(buildPartial()); 4438 } 4439 4440 public com.google.protobuf.Descriptors.Descriptor 4441 getDescriptorForType() { 4442 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor; 4443 } 4444 4445 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto getDefaultInstanceForType() { 4446 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.getDefaultInstance(); 4447 } 4448 4449 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto build() { 4450 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto result = buildPartial(); 4451 if (!result.isInitialized()) { 4452 throw newUninitializedMessageException(result); 4453 } 4454 return result; 4455 } 4456 4457 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto buildPartial() { 4458 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto(this); 4459 int from_bitField0_ = bitField0_; 4460 int to_bitField0_ = 0; 4461 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 4462 to_bitField0_ |= 0x00000001; 4463 } 4464 result.name_ = name_; 4465 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 4466 to_bitField0_ |= 0x00000002; 4467 } 4468 result.value_ = value_; 4469 result.bitField0_ = to_bitField0_; 4470 onBuilt(); 4471 return result; 4472 } 4473 4474 public Builder mergeFrom(com.google.protobuf.Message other) { 4475 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto) { 4476 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto)other); 4477 } else { 4478 super.mergeFrom(other); 4479 return this; 4480 } 4481 } 4482 4483 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto other) { 4484 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.getDefaultInstance()) return this; 4485 if (other.hasName()) { 4486 setName(other.getName()); 4487 } 4488 if (other.hasValue()) { 4489 setValue(other.getValue()); 4490 } 4491 this.mergeUnknownFields(other.getUnknownFields()); 4492 return this; 4493 } 4494 4495 public final boolean isInitialized() { 4496 if (!hasName()) { 4497 4498 return false; 4499 } 4500 return true; 4501 } 4502 4503 public Builder mergeFrom( 4504 com.google.protobuf.CodedInputStream input, 4505 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4506 throws java.io.IOException { 4507 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto parsedMessage = null; 4508 try { 4509 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4510 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4511 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto) e.getUnfinishedMessage(); 4512 throw e; 4513 } finally { 4514 if (parsedMessage != null) { 4515 mergeFrom(parsedMessage); 4516 } 4517 } 4518 return this; 4519 } 4520 private int bitField0_; 4521 4522 // required fixed32 name = 1; 4523 private int name_ ; 4524 /** 4525 * <code>required fixed32 name = 1;</code> 4526 * 4527 * <pre> 4528 ** 4529 * 4530 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4531 * [2:26) -- the name of the entry, which is an ID that points to a 4532 * string in the StringTableSection. 4533 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4534 * so only 2 bits were needed. At that time, this bit was reserved. When a 4535 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4536 * bit. 4537 * [27:32) -- reserved for future uses. 4538 * </pre> 4539 */ 4540 public boolean hasName() { 4541 return ((bitField0_ & 0x00000001) == 0x00000001); 4542 } 4543 /** 4544 * <code>required fixed32 name = 1;</code> 4545 * 4546 * <pre> 4547 ** 4548 * 4549 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4550 * [2:26) -- the name of the entry, which is an ID that points to a 4551 * string in the StringTableSection. 4552 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4553 * so only 2 bits were needed. At that time, this bit was reserved. When a 4554 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4555 * bit. 4556 * [27:32) -- reserved for future uses. 4557 * </pre> 4558 */ 4559 public int getName() { 4560 return name_; 4561 } 4562 /** 4563 * <code>required fixed32 name = 1;</code> 4564 * 4565 * <pre> 4566 ** 4567 * 4568 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4569 * [2:26) -- the name of the entry, which is an ID that points to a 4570 * string in the StringTableSection. 4571 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4572 * so only 2 bits were needed. At that time, this bit was reserved. When a 4573 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4574 * bit. 4575 * [27:32) -- reserved for future uses. 4576 * </pre> 4577 */ 4578 public Builder setName(int value) { 4579 bitField0_ |= 0x00000001; 4580 name_ = value; 4581 onChanged(); 4582 return this; 4583 } 4584 /** 4585 * <code>required fixed32 name = 1;</code> 4586 * 4587 * <pre> 4588 ** 4589 * 4590 * [0:2) -- the namespace of XAttr (XAttrNamespaceProto) 4591 * [2:26) -- the name of the entry, which is an ID that points to a 4592 * string in the StringTableSection. 4593 * [26:27) -- namespace extension. Originally there were only 4 namespaces 4594 * so only 2 bits were needed. At that time, this bit was reserved. When a 4595 * 5th namespace was created (raw) this bit became used as a 3rd namespace 4596 * bit. 4597 * [27:32) -- reserved for future uses. 4598 * </pre> 4599 */ 4600 public Builder clearName() { 4601 bitField0_ = (bitField0_ & ~0x00000001); 4602 name_ = 0; 4603 onChanged(); 4604 return this; 4605 } 4606 4607 // optional bytes value = 2; 4608 private com.google.protobuf.ByteString value_ = com.google.protobuf.ByteString.EMPTY; 4609 /** 4610 * <code>optional bytes value = 2;</code> 4611 */ 4612 public boolean hasValue() { 4613 return ((bitField0_ & 0x00000002) == 0x00000002); 4614 } 4615 /** 4616 * <code>optional bytes value = 2;</code> 4617 */ 4618 public com.google.protobuf.ByteString getValue() { 4619 return value_; 4620 } 4621 /** 4622 * <code>optional bytes value = 2;</code> 4623 */ 4624 public Builder setValue(com.google.protobuf.ByteString value) { 4625 if (value == null) { 4626 throw new NullPointerException(); 4627 } 4628 bitField0_ |= 0x00000002; 4629 value_ = value; 4630 onChanged(); 4631 return this; 4632 } 4633 /** 4634 * <code>optional bytes value = 2;</code> 4635 */ 4636 public Builder clearValue() { 4637 bitField0_ = (bitField0_ & ~0x00000002); 4638 value_ = getDefaultInstance().getValue(); 4639 onChanged(); 4640 return this; 4641 } 4642 4643 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto) 4644 } 4645 4646 static { 4647 defaultInstance = new XAttrCompactProto(true); 4648 defaultInstance.initFields(); 4649 } 4650 4651 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto) 4652 } 4653 4654 public interface XAttrFeatureProtoOrBuilder 4655 extends com.google.protobuf.MessageOrBuilder { 4656 4657 // repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1; 4658 /** 4659 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4660 */ 4661 java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> 4662 getXAttrsList(); 4663 /** 4664 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4665 */ 4666 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto getXAttrs(int index); 4667 /** 4668 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4669 */ 4670 int getXAttrsCount(); 4671 /** 4672 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4673 */ 4674 java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder> 4675 getXAttrsOrBuilderList(); 4676 /** 4677 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4678 */ 4679 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder getXAttrsOrBuilder( 4680 int index); 4681 } 4682 /** 4683 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto} 4684 */ 4685 public static final class XAttrFeatureProto extends 4686 com.google.protobuf.GeneratedMessage 4687 implements XAttrFeatureProtoOrBuilder { 4688 // Use XAttrFeatureProto.newBuilder() to construct. 4689 private XAttrFeatureProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4690 super(builder); 4691 this.unknownFields = builder.getUnknownFields(); 4692 } 4693 private XAttrFeatureProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4694 4695 private static final XAttrFeatureProto defaultInstance; 4696 public static XAttrFeatureProto getDefaultInstance() { 4697 return defaultInstance; 4698 } 4699 4700 public XAttrFeatureProto getDefaultInstanceForType() { 4701 return defaultInstance; 4702 } 4703 4704 private final com.google.protobuf.UnknownFieldSet unknownFields; 4705 @java.lang.Override 4706 public final com.google.protobuf.UnknownFieldSet 4707 getUnknownFields() { 4708 return this.unknownFields; 4709 } 4710 private XAttrFeatureProto( 4711 com.google.protobuf.CodedInputStream input, 4712 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4713 throws com.google.protobuf.InvalidProtocolBufferException { 4714 initFields(); 4715 int mutable_bitField0_ = 0; 4716 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4717 com.google.protobuf.UnknownFieldSet.newBuilder(); 4718 try { 4719 boolean done = false; 4720 while (!done) { 4721 int tag = input.readTag(); 4722 switch (tag) { 4723 case 0: 4724 done = true; 4725 break; 4726 default: { 4727 if (!parseUnknownField(input, unknownFields, 4728 extensionRegistry, tag)) { 4729 done = true; 4730 } 4731 break; 4732 } 4733 case 10: { 4734 if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 4735 xAttrs_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto>(); 4736 mutable_bitField0_ |= 0x00000001; 4737 } 4738 xAttrs_.add(input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.PARSER, extensionRegistry)); 4739 break; 4740 } 4741 } 4742 } 4743 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4744 throw e.setUnfinishedMessage(this); 4745 } catch (java.io.IOException e) { 4746 throw new com.google.protobuf.InvalidProtocolBufferException( 4747 e.getMessage()).setUnfinishedMessage(this); 4748 } finally { 4749 if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { 4750 xAttrs_ = java.util.Collections.unmodifiableList(xAttrs_); 4751 } 4752 this.unknownFields = unknownFields.build(); 4753 makeExtensionsImmutable(); 4754 } 4755 } 4756 public static final com.google.protobuf.Descriptors.Descriptor 4757 getDescriptor() { 4758 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor; 4759 } 4760 4761 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4762 internalGetFieldAccessorTable() { 4763 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_fieldAccessorTable 4764 .ensureFieldAccessorsInitialized( 4765 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder.class); 4766 } 4767 4768 public static com.google.protobuf.Parser<XAttrFeatureProto> PARSER = 4769 new com.google.protobuf.AbstractParser<XAttrFeatureProto>() { 4770 public XAttrFeatureProto parsePartialFrom( 4771 com.google.protobuf.CodedInputStream input, 4772 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4773 throws com.google.protobuf.InvalidProtocolBufferException { 4774 return new XAttrFeatureProto(input, extensionRegistry); 4775 } 4776 }; 4777 4778 @java.lang.Override 4779 public com.google.protobuf.Parser<XAttrFeatureProto> getParserForType() { 4780 return PARSER; 4781 } 4782 4783 // repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1; 4784 public static final int XATTRS_FIELD_NUMBER = 1; 4785 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> xAttrs_; 4786 /** 4787 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4788 */ 4789 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> getXAttrsList() { 4790 return xAttrs_; 4791 } 4792 /** 4793 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4794 */ 4795 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder> 4796 getXAttrsOrBuilderList() { 4797 return xAttrs_; 4798 } 4799 /** 4800 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4801 */ 4802 public int getXAttrsCount() { 4803 return xAttrs_.size(); 4804 } 4805 /** 4806 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4807 */ 4808 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto getXAttrs(int index) { 4809 return xAttrs_.get(index); 4810 } 4811 /** 4812 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 4813 */ 4814 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder getXAttrsOrBuilder( 4815 int index) { 4816 return xAttrs_.get(index); 4817 } 4818 4819 private void initFields() { 4820 xAttrs_ = java.util.Collections.emptyList(); 4821 } 4822 private byte memoizedIsInitialized = -1; 4823 public final boolean isInitialized() { 4824 byte isInitialized = memoizedIsInitialized; 4825 if (isInitialized != -1) return isInitialized == 1; 4826 4827 for (int i = 0; i < getXAttrsCount(); i++) { 4828 if (!getXAttrs(i).isInitialized()) { 4829 memoizedIsInitialized = 0; 4830 return false; 4831 } 4832 } 4833 memoizedIsInitialized = 1; 4834 return true; 4835 } 4836 4837 public void writeTo(com.google.protobuf.CodedOutputStream output) 4838 throws java.io.IOException { 4839 getSerializedSize(); 4840 for (int i = 0; i < xAttrs_.size(); i++) { 4841 output.writeMessage(1, xAttrs_.get(i)); 4842 } 4843 getUnknownFields().writeTo(output); 4844 } 4845 4846 private int memoizedSerializedSize = -1; 4847 public int getSerializedSize() { 4848 int size = memoizedSerializedSize; 4849 if (size != -1) return size; 4850 4851 size = 0; 4852 for (int i = 0; i < xAttrs_.size(); i++) { 4853 size += com.google.protobuf.CodedOutputStream 4854 .computeMessageSize(1, xAttrs_.get(i)); 4855 } 4856 size += getUnknownFields().getSerializedSize(); 4857 memoizedSerializedSize = size; 4858 return size; 4859 } 4860 4861 private static final long serialVersionUID = 0L; 4862 @java.lang.Override 4863 protected java.lang.Object writeReplace() 4864 throws java.io.ObjectStreamException { 4865 return super.writeReplace(); 4866 } 4867 4868 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4869 com.google.protobuf.ByteString data) 4870 throws com.google.protobuf.InvalidProtocolBufferException { 4871 return PARSER.parseFrom(data); 4872 } 4873 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4874 com.google.protobuf.ByteString data, 4875 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4876 throws com.google.protobuf.InvalidProtocolBufferException { 4877 return PARSER.parseFrom(data, extensionRegistry); 4878 } 4879 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom(byte[] data) 4880 throws com.google.protobuf.InvalidProtocolBufferException { 4881 return PARSER.parseFrom(data); 4882 } 4883 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4884 byte[] data, 4885 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4886 throws com.google.protobuf.InvalidProtocolBufferException { 4887 return PARSER.parseFrom(data, extensionRegistry); 4888 } 4889 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom(java.io.InputStream input) 4890 throws java.io.IOException { 4891 return PARSER.parseFrom(input); 4892 } 4893 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4894 java.io.InputStream input, 4895 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4896 throws java.io.IOException { 4897 return PARSER.parseFrom(input, extensionRegistry); 4898 } 4899 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseDelimitedFrom(java.io.InputStream input) 4900 throws java.io.IOException { 4901 return PARSER.parseDelimitedFrom(input); 4902 } 4903 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseDelimitedFrom( 4904 java.io.InputStream input, 4905 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4906 throws java.io.IOException { 4907 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4908 } 4909 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4910 com.google.protobuf.CodedInputStream input) 4911 throws java.io.IOException { 4912 return PARSER.parseFrom(input); 4913 } 4914 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parseFrom( 4915 com.google.protobuf.CodedInputStream input, 4916 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4917 throws java.io.IOException { 4918 return PARSER.parseFrom(input, extensionRegistry); 4919 } 4920 4921 public static Builder newBuilder() { return Builder.create(); } 4922 public Builder newBuilderForType() { return newBuilder(); } 4923 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto prototype) { 4924 return newBuilder().mergeFrom(prototype); 4925 } 4926 public Builder toBuilder() { return newBuilder(this); } 4927 4928 @java.lang.Override 4929 protected Builder newBuilderForType( 4930 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4931 Builder builder = new Builder(parent); 4932 return builder; 4933 } 4934 /** 4935 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto} 4936 */ 4937 public static final class Builder extends 4938 com.google.protobuf.GeneratedMessage.Builder<Builder> 4939 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder { 4940 public static final com.google.protobuf.Descriptors.Descriptor 4941 getDescriptor() { 4942 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor; 4943 } 4944 4945 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4946 internalGetFieldAccessorTable() { 4947 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_fieldAccessorTable 4948 .ensureFieldAccessorsInitialized( 4949 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder.class); 4950 } 4951 4952 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.newBuilder() 4953 private Builder() { 4954 maybeForceBuilderInitialization(); 4955 } 4956 4957 private Builder( 4958 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4959 super(parent); 4960 maybeForceBuilderInitialization(); 4961 } 4962 private void maybeForceBuilderInitialization() { 4963 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4964 getXAttrsFieldBuilder(); 4965 } 4966 } 4967 private static Builder create() { 4968 return new Builder(); 4969 } 4970 4971 public Builder clear() { 4972 super.clear(); 4973 if (xAttrsBuilder_ == null) { 4974 xAttrs_ = java.util.Collections.emptyList(); 4975 bitField0_ = (bitField0_ & ~0x00000001); 4976 } else { 4977 xAttrsBuilder_.clear(); 4978 } 4979 return this; 4980 } 4981 4982 public Builder clone() { 4983 return create().mergeFrom(buildPartial()); 4984 } 4985 4986 public com.google.protobuf.Descriptors.Descriptor 4987 getDescriptorForType() { 4988 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor; 4989 } 4990 4991 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getDefaultInstanceForType() { 4992 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 4993 } 4994 4995 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto build() { 4996 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto result = buildPartial(); 4997 if (!result.isInitialized()) { 4998 throw newUninitializedMessageException(result); 4999 } 5000 return result; 5001 } 5002 5003 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto buildPartial() { 5004 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto(this); 5005 int from_bitField0_ = bitField0_; 5006 if (xAttrsBuilder_ == null) { 5007 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5008 xAttrs_ = java.util.Collections.unmodifiableList(xAttrs_); 5009 bitField0_ = (bitField0_ & ~0x00000001); 5010 } 5011 result.xAttrs_ = xAttrs_; 5012 } else { 5013 result.xAttrs_ = xAttrsBuilder_.build(); 5014 } 5015 onBuilt(); 5016 return result; 5017 } 5018 5019 public Builder mergeFrom(com.google.protobuf.Message other) { 5020 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto) { 5021 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto)other); 5022 } else { 5023 super.mergeFrom(other); 5024 return this; 5025 } 5026 } 5027 5028 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto other) { 5029 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance()) return this; 5030 if (xAttrsBuilder_ == null) { 5031 if (!other.xAttrs_.isEmpty()) { 5032 if (xAttrs_.isEmpty()) { 5033 xAttrs_ = other.xAttrs_; 5034 bitField0_ = (bitField0_ & ~0x00000001); 5035 } else { 5036 ensureXAttrsIsMutable(); 5037 xAttrs_.addAll(other.xAttrs_); 5038 } 5039 onChanged(); 5040 } 5041 } else { 5042 if (!other.xAttrs_.isEmpty()) { 5043 if (xAttrsBuilder_.isEmpty()) { 5044 xAttrsBuilder_.dispose(); 5045 xAttrsBuilder_ = null; 5046 xAttrs_ = other.xAttrs_; 5047 bitField0_ = (bitField0_ & ~0x00000001); 5048 xAttrsBuilder_ = 5049 com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? 5050 getXAttrsFieldBuilder() : null; 5051 } else { 5052 xAttrsBuilder_.addAllMessages(other.xAttrs_); 5053 } 5054 } 5055 } 5056 this.mergeUnknownFields(other.getUnknownFields()); 5057 return this; 5058 } 5059 5060 public final boolean isInitialized() { 5061 for (int i = 0; i < getXAttrsCount(); i++) { 5062 if (!getXAttrs(i).isInitialized()) { 5063 5064 return false; 5065 } 5066 } 5067 return true; 5068 } 5069 5070 public Builder mergeFrom( 5071 com.google.protobuf.CodedInputStream input, 5072 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5073 throws java.io.IOException { 5074 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto parsedMessage = null; 5075 try { 5076 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 5077 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5078 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto) e.getUnfinishedMessage(); 5079 throw e; 5080 } finally { 5081 if (parsedMessage != null) { 5082 mergeFrom(parsedMessage); 5083 } 5084 } 5085 return this; 5086 } 5087 private int bitField0_; 5088 5089 // repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1; 5090 private java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> xAttrs_ = 5091 java.util.Collections.emptyList(); 5092 private void ensureXAttrsIsMutable() { 5093 if (!((bitField0_ & 0x00000001) == 0x00000001)) { 5094 xAttrs_ = new java.util.ArrayList<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto>(xAttrs_); 5095 bitField0_ |= 0x00000001; 5096 } 5097 } 5098 5099 private com.google.protobuf.RepeatedFieldBuilder< 5100 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder> xAttrsBuilder_; 5101 5102 /** 5103 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5104 */ 5105 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> getXAttrsList() { 5106 if (xAttrsBuilder_ == null) { 5107 return java.util.Collections.unmodifiableList(xAttrs_); 5108 } else { 5109 return xAttrsBuilder_.getMessageList(); 5110 } 5111 } 5112 /** 5113 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5114 */ 5115 public int getXAttrsCount() { 5116 if (xAttrsBuilder_ == null) { 5117 return xAttrs_.size(); 5118 } else { 5119 return xAttrsBuilder_.getCount(); 5120 } 5121 } 5122 /** 5123 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5124 */ 5125 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto getXAttrs(int index) { 5126 if (xAttrsBuilder_ == null) { 5127 return xAttrs_.get(index); 5128 } else { 5129 return xAttrsBuilder_.getMessage(index); 5130 } 5131 } 5132 /** 5133 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5134 */ 5135 public Builder setXAttrs( 5136 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto value) { 5137 if (xAttrsBuilder_ == null) { 5138 if (value == null) { 5139 throw new NullPointerException(); 5140 } 5141 ensureXAttrsIsMutable(); 5142 xAttrs_.set(index, value); 5143 onChanged(); 5144 } else { 5145 xAttrsBuilder_.setMessage(index, value); 5146 } 5147 return this; 5148 } 5149 /** 5150 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5151 */ 5152 public Builder setXAttrs( 5153 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder builderForValue) { 5154 if (xAttrsBuilder_ == null) { 5155 ensureXAttrsIsMutable(); 5156 xAttrs_.set(index, builderForValue.build()); 5157 onChanged(); 5158 } else { 5159 xAttrsBuilder_.setMessage(index, builderForValue.build()); 5160 } 5161 return this; 5162 } 5163 /** 5164 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5165 */ 5166 public Builder addXAttrs(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto value) { 5167 if (xAttrsBuilder_ == null) { 5168 if (value == null) { 5169 throw new NullPointerException(); 5170 } 5171 ensureXAttrsIsMutable(); 5172 xAttrs_.add(value); 5173 onChanged(); 5174 } else { 5175 xAttrsBuilder_.addMessage(value); 5176 } 5177 return this; 5178 } 5179 /** 5180 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5181 */ 5182 public Builder addXAttrs( 5183 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto value) { 5184 if (xAttrsBuilder_ == null) { 5185 if (value == null) { 5186 throw new NullPointerException(); 5187 } 5188 ensureXAttrsIsMutable(); 5189 xAttrs_.add(index, value); 5190 onChanged(); 5191 } else { 5192 xAttrsBuilder_.addMessage(index, value); 5193 } 5194 return this; 5195 } 5196 /** 5197 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5198 */ 5199 public Builder addXAttrs( 5200 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder builderForValue) { 5201 if (xAttrsBuilder_ == null) { 5202 ensureXAttrsIsMutable(); 5203 xAttrs_.add(builderForValue.build()); 5204 onChanged(); 5205 } else { 5206 xAttrsBuilder_.addMessage(builderForValue.build()); 5207 } 5208 return this; 5209 } 5210 /** 5211 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5212 */ 5213 public Builder addXAttrs( 5214 int index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder builderForValue) { 5215 if (xAttrsBuilder_ == null) { 5216 ensureXAttrsIsMutable(); 5217 xAttrs_.add(index, builderForValue.build()); 5218 onChanged(); 5219 } else { 5220 xAttrsBuilder_.addMessage(index, builderForValue.build()); 5221 } 5222 return this; 5223 } 5224 /** 5225 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5226 */ 5227 public Builder addAllXAttrs( 5228 java.lang.Iterable<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto> values) { 5229 if (xAttrsBuilder_ == null) { 5230 ensureXAttrsIsMutable(); 5231 super.addAll(values, xAttrs_); 5232 onChanged(); 5233 } else { 5234 xAttrsBuilder_.addAllMessages(values); 5235 } 5236 return this; 5237 } 5238 /** 5239 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5240 */ 5241 public Builder clearXAttrs() { 5242 if (xAttrsBuilder_ == null) { 5243 xAttrs_ = java.util.Collections.emptyList(); 5244 bitField0_ = (bitField0_ & ~0x00000001); 5245 onChanged(); 5246 } else { 5247 xAttrsBuilder_.clear(); 5248 } 5249 return this; 5250 } 5251 /** 5252 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5253 */ 5254 public Builder removeXAttrs(int index) { 5255 if (xAttrsBuilder_ == null) { 5256 ensureXAttrsIsMutable(); 5257 xAttrs_.remove(index); 5258 onChanged(); 5259 } else { 5260 xAttrsBuilder_.remove(index); 5261 } 5262 return this; 5263 } 5264 /** 5265 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5266 */ 5267 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder getXAttrsBuilder( 5268 int index) { 5269 return getXAttrsFieldBuilder().getBuilder(index); 5270 } 5271 /** 5272 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5273 */ 5274 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder getXAttrsOrBuilder( 5275 int index) { 5276 if (xAttrsBuilder_ == null) { 5277 return xAttrs_.get(index); } else { 5278 return xAttrsBuilder_.getMessageOrBuilder(index); 5279 } 5280 } 5281 /** 5282 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5283 */ 5284 public java.util.List<? extends org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder> 5285 getXAttrsOrBuilderList() { 5286 if (xAttrsBuilder_ != null) { 5287 return xAttrsBuilder_.getMessageOrBuilderList(); 5288 } else { 5289 return java.util.Collections.unmodifiableList(xAttrs_); 5290 } 5291 } 5292 /** 5293 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5294 */ 5295 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder addXAttrsBuilder() { 5296 return getXAttrsFieldBuilder().addBuilder( 5297 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.getDefaultInstance()); 5298 } 5299 /** 5300 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5301 */ 5302 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder addXAttrsBuilder( 5303 int index) { 5304 return getXAttrsFieldBuilder().addBuilder( 5305 index, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.getDefaultInstance()); 5306 } 5307 /** 5308 * <code>repeated .hadoop.hdfs.fsimage.INodeSection.XAttrCompactProto xAttrs = 1;</code> 5309 */ 5310 public java.util.List<org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder> 5311 getXAttrsBuilderList() { 5312 return getXAttrsFieldBuilder().getBuilderList(); 5313 } 5314 private com.google.protobuf.RepeatedFieldBuilder< 5315 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder> 5316 getXAttrsFieldBuilder() { 5317 if (xAttrsBuilder_ == null) { 5318 xAttrsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< 5319 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrCompactProtoOrBuilder>( 5320 xAttrs_, 5321 ((bitField0_ & 0x00000001) == 0x00000001), 5322 getParentForChildren(), 5323 isClean()); 5324 xAttrs_ = null; 5325 } 5326 return xAttrsBuilder_; 5327 } 5328 5329 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto) 5330 } 5331 5332 static { 5333 defaultInstance = new XAttrFeatureProto(true); 5334 defaultInstance.initFields(); 5335 } 5336 5337 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto) 5338 } 5339 5340 public interface INodeFileOrBuilder 5341 extends com.google.protobuf.MessageOrBuilder { 5342 5343 // optional uint32 replication = 1; 5344 /** 5345 * <code>optional uint32 replication = 1;</code> 5346 */ 5347 boolean hasReplication(); 5348 /** 5349 * <code>optional uint32 replication = 1;</code> 5350 */ 5351 int getReplication(); 5352 5353 // optional uint64 modificationTime = 2; 5354 /** 5355 * <code>optional uint64 modificationTime = 2;</code> 5356 */ 5357 boolean hasModificationTime(); 5358 /** 5359 * <code>optional uint64 modificationTime = 2;</code> 5360 */ 5361 long getModificationTime(); 5362 5363 // optional uint64 accessTime = 3; 5364 /** 5365 * <code>optional uint64 accessTime = 3;</code> 5366 */ 5367 boolean hasAccessTime(); 5368 /** 5369 * <code>optional uint64 accessTime = 3;</code> 5370 */ 5371 long getAccessTime(); 5372 5373 // optional uint64 preferredBlockSize = 4; 5374 /** 5375 * <code>optional uint64 preferredBlockSize = 4;</code> 5376 */ 5377 boolean hasPreferredBlockSize(); 5378 /** 5379 * <code>optional uint64 preferredBlockSize = 4;</code> 5380 */ 5381 long getPreferredBlockSize(); 5382 5383 // optional fixed64 permission = 5; 5384 /** 5385 * <code>optional fixed64 permission = 5;</code> 5386 */ 5387 boolean hasPermission(); 5388 /** 5389 * <code>optional fixed64 permission = 5;</code> 5390 */ 5391 long getPermission(); 5392 5393 // repeated .hadoop.hdfs.BlockProto blocks = 6; 5394 /** 5395 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5396 */ 5397 java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> 5398 getBlocksList(); 5399 /** 5400 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5401 */ 5402 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index); 5403 /** 5404 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5405 */ 5406 int getBlocksCount(); 5407 /** 5408 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5409 */ 5410 java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 5411 getBlocksOrBuilderList(); 5412 /** 5413 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5414 */ 5415 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 5416 int index); 5417 5418 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 5419 /** 5420 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5421 */ 5422 boolean hasFileUC(); 5423 /** 5424 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5425 */ 5426 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC(); 5427 /** 5428 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5429 */ 5430 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder(); 5431 5432 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 5433 /** 5434 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5435 */ 5436 boolean hasAcl(); 5437 /** 5438 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5439 */ 5440 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl(); 5441 /** 5442 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5443 */ 5444 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder(); 5445 5446 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9; 5447 /** 5448 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5449 */ 5450 boolean hasXAttrs(); 5451 /** 5452 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5453 */ 5454 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs(); 5455 /** 5456 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5457 */ 5458 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder(); 5459 5460 // optional uint32 storagePolicyID = 10; 5461 /** 5462 * <code>optional uint32 storagePolicyID = 10;</code> 5463 */ 5464 boolean hasStoragePolicyID(); 5465 /** 5466 * <code>optional uint32 storagePolicyID = 10;</code> 5467 */ 5468 int getStoragePolicyID(); 5469 } 5470 /** 5471 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeFile} 5472 */ 5473 public static final class INodeFile extends 5474 com.google.protobuf.GeneratedMessage 5475 implements INodeFileOrBuilder { 5476 // Use INodeFile.newBuilder() to construct. 5477 private INodeFile(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 5478 super(builder); 5479 this.unknownFields = builder.getUnknownFields(); 5480 } 5481 private INodeFile(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 5482 5483 private static final INodeFile defaultInstance; 5484 public static INodeFile getDefaultInstance() { 5485 return defaultInstance; 5486 } 5487 5488 public INodeFile getDefaultInstanceForType() { 5489 return defaultInstance; 5490 } 5491 5492 private final com.google.protobuf.UnknownFieldSet unknownFields; 5493 @java.lang.Override 5494 public final com.google.protobuf.UnknownFieldSet 5495 getUnknownFields() { 5496 return this.unknownFields; 5497 } 5498 private INodeFile( 5499 com.google.protobuf.CodedInputStream input, 5500 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5501 throws com.google.protobuf.InvalidProtocolBufferException { 5502 initFields(); 5503 int mutable_bitField0_ = 0; 5504 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 5505 com.google.protobuf.UnknownFieldSet.newBuilder(); 5506 try { 5507 boolean done = false; 5508 while (!done) { 5509 int tag = input.readTag(); 5510 switch (tag) { 5511 case 0: 5512 done = true; 5513 break; 5514 default: { 5515 if (!parseUnknownField(input, unknownFields, 5516 extensionRegistry, tag)) { 5517 done = true; 5518 } 5519 break; 5520 } 5521 case 8: { 5522 bitField0_ |= 0x00000001; 5523 replication_ = input.readUInt32(); 5524 break; 5525 } 5526 case 16: { 5527 bitField0_ |= 0x00000002; 5528 modificationTime_ = input.readUInt64(); 5529 break; 5530 } 5531 case 24: { 5532 bitField0_ |= 0x00000004; 5533 accessTime_ = input.readUInt64(); 5534 break; 5535 } 5536 case 32: { 5537 bitField0_ |= 0x00000008; 5538 preferredBlockSize_ = input.readUInt64(); 5539 break; 5540 } 5541 case 41: { 5542 bitField0_ |= 0x00000010; 5543 permission_ = input.readFixed64(); 5544 break; 5545 } 5546 case 50: { 5547 if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) { 5548 blocks_ = new java.util.ArrayList<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto>(); 5549 mutable_bitField0_ |= 0x00000020; 5550 } 5551 blocks_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.PARSER, extensionRegistry)); 5552 break; 5553 } 5554 case 58: { 5555 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder subBuilder = null; 5556 if (((bitField0_ & 0x00000020) == 0x00000020)) { 5557 subBuilder = fileUC_.toBuilder(); 5558 } 5559 fileUC_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.PARSER, extensionRegistry); 5560 if (subBuilder != null) { 5561 subBuilder.mergeFrom(fileUC_); 5562 fileUC_ = subBuilder.buildPartial(); 5563 } 5564 bitField0_ |= 0x00000020; 5565 break; 5566 } 5567 case 66: { 5568 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder subBuilder = null; 5569 if (((bitField0_ & 0x00000040) == 0x00000040)) { 5570 subBuilder = acl_.toBuilder(); 5571 } 5572 acl_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.PARSER, extensionRegistry); 5573 if (subBuilder != null) { 5574 subBuilder.mergeFrom(acl_); 5575 acl_ = subBuilder.buildPartial(); 5576 } 5577 bitField0_ |= 0x00000040; 5578 break; 5579 } 5580 case 74: { 5581 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder subBuilder = null; 5582 if (((bitField0_ & 0x00000080) == 0x00000080)) { 5583 subBuilder = xAttrs_.toBuilder(); 5584 } 5585 xAttrs_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.PARSER, extensionRegistry); 5586 if (subBuilder != null) { 5587 subBuilder.mergeFrom(xAttrs_); 5588 xAttrs_ = subBuilder.buildPartial(); 5589 } 5590 bitField0_ |= 0x00000080; 5591 break; 5592 } 5593 case 80: { 5594 bitField0_ |= 0x00000100; 5595 storagePolicyID_ = input.readUInt32(); 5596 break; 5597 } 5598 } 5599 } 5600 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5601 throw e.setUnfinishedMessage(this); 5602 } catch (java.io.IOException e) { 5603 throw new com.google.protobuf.InvalidProtocolBufferException( 5604 e.getMessage()).setUnfinishedMessage(this); 5605 } finally { 5606 if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) { 5607 blocks_ = java.util.Collections.unmodifiableList(blocks_); 5608 } 5609 this.unknownFields = unknownFields.build(); 5610 makeExtensionsImmutable(); 5611 } 5612 } 5613 public static final com.google.protobuf.Descriptors.Descriptor 5614 getDescriptor() { 5615 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 5616 } 5617 5618 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 5619 internalGetFieldAccessorTable() { 5620 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable 5621 .ensureFieldAccessorsInitialized( 5622 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder.class); 5623 } 5624 5625 public static com.google.protobuf.Parser<INodeFile> PARSER = 5626 new com.google.protobuf.AbstractParser<INodeFile>() { 5627 public INodeFile parsePartialFrom( 5628 com.google.protobuf.CodedInputStream input, 5629 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5630 throws com.google.protobuf.InvalidProtocolBufferException { 5631 return new INodeFile(input, extensionRegistry); 5632 } 5633 }; 5634 5635 @java.lang.Override 5636 public com.google.protobuf.Parser<INodeFile> getParserForType() { 5637 return PARSER; 5638 } 5639 5640 private int bitField0_; 5641 // optional uint32 replication = 1; 5642 public static final int REPLICATION_FIELD_NUMBER = 1; 5643 private int replication_; 5644 /** 5645 * <code>optional uint32 replication = 1;</code> 5646 */ 5647 public boolean hasReplication() { 5648 return ((bitField0_ & 0x00000001) == 0x00000001); 5649 } 5650 /** 5651 * <code>optional uint32 replication = 1;</code> 5652 */ 5653 public int getReplication() { 5654 return replication_; 5655 } 5656 5657 // optional uint64 modificationTime = 2; 5658 public static final int MODIFICATIONTIME_FIELD_NUMBER = 2; 5659 private long modificationTime_; 5660 /** 5661 * <code>optional uint64 modificationTime = 2;</code> 5662 */ 5663 public boolean hasModificationTime() { 5664 return ((bitField0_ & 0x00000002) == 0x00000002); 5665 } 5666 /** 5667 * <code>optional uint64 modificationTime = 2;</code> 5668 */ 5669 public long getModificationTime() { 5670 return modificationTime_; 5671 } 5672 5673 // optional uint64 accessTime = 3; 5674 public static final int ACCESSTIME_FIELD_NUMBER = 3; 5675 private long accessTime_; 5676 /** 5677 * <code>optional uint64 accessTime = 3;</code> 5678 */ 5679 public boolean hasAccessTime() { 5680 return ((bitField0_ & 0x00000004) == 0x00000004); 5681 } 5682 /** 5683 * <code>optional uint64 accessTime = 3;</code> 5684 */ 5685 public long getAccessTime() { 5686 return accessTime_; 5687 } 5688 5689 // optional uint64 preferredBlockSize = 4; 5690 public static final int PREFERREDBLOCKSIZE_FIELD_NUMBER = 4; 5691 private long preferredBlockSize_; 5692 /** 5693 * <code>optional uint64 preferredBlockSize = 4;</code> 5694 */ 5695 public boolean hasPreferredBlockSize() { 5696 return ((bitField0_ & 0x00000008) == 0x00000008); 5697 } 5698 /** 5699 * <code>optional uint64 preferredBlockSize = 4;</code> 5700 */ 5701 public long getPreferredBlockSize() { 5702 return preferredBlockSize_; 5703 } 5704 5705 // optional fixed64 permission = 5; 5706 public static final int PERMISSION_FIELD_NUMBER = 5; 5707 private long permission_; 5708 /** 5709 * <code>optional fixed64 permission = 5;</code> 5710 */ 5711 public boolean hasPermission() { 5712 return ((bitField0_ & 0x00000010) == 0x00000010); 5713 } 5714 /** 5715 * <code>optional fixed64 permission = 5;</code> 5716 */ 5717 public long getPermission() { 5718 return permission_; 5719 } 5720 5721 // repeated .hadoop.hdfs.BlockProto blocks = 6; 5722 public static final int BLOCKS_FIELD_NUMBER = 6; 5723 private java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> blocks_; 5724 /** 5725 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5726 */ 5727 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> getBlocksList() { 5728 return blocks_; 5729 } 5730 /** 5731 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5732 */ 5733 public java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 5734 getBlocksOrBuilderList() { 5735 return blocks_; 5736 } 5737 /** 5738 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5739 */ 5740 public int getBlocksCount() { 5741 return blocks_.size(); 5742 } 5743 /** 5744 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5745 */ 5746 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { 5747 return blocks_.get(index); 5748 } 5749 /** 5750 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 5751 */ 5752 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 5753 int index) { 5754 return blocks_.get(index); 5755 } 5756 5757 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 5758 public static final int FILEUC_FIELD_NUMBER = 7; 5759 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature fileUC_; 5760 /** 5761 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5762 */ 5763 public boolean hasFileUC() { 5764 return ((bitField0_ & 0x00000020) == 0x00000020); 5765 } 5766 /** 5767 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5768 */ 5769 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC() { 5770 return fileUC_; 5771 } 5772 /** 5773 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 5774 */ 5775 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder() { 5776 return fileUC_; 5777 } 5778 5779 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 5780 public static final int ACL_FIELD_NUMBER = 8; 5781 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_; 5782 /** 5783 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5784 */ 5785 public boolean hasAcl() { 5786 return ((bitField0_ & 0x00000040) == 0x00000040); 5787 } 5788 /** 5789 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5790 */ 5791 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 5792 return acl_; 5793 } 5794 /** 5795 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 5796 */ 5797 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 5798 return acl_; 5799 } 5800 5801 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9; 5802 public static final int XATTRS_FIELD_NUMBER = 9; 5803 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto xAttrs_; 5804 /** 5805 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5806 */ 5807 public boolean hasXAttrs() { 5808 return ((bitField0_ & 0x00000080) == 0x00000080); 5809 } 5810 /** 5811 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5812 */ 5813 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs() { 5814 return xAttrs_; 5815 } 5816 /** 5817 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 5818 */ 5819 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder() { 5820 return xAttrs_; 5821 } 5822 5823 // optional uint32 storagePolicyID = 10; 5824 public static final int STORAGEPOLICYID_FIELD_NUMBER = 10; 5825 private int storagePolicyID_; 5826 /** 5827 * <code>optional uint32 storagePolicyID = 10;</code> 5828 */ 5829 public boolean hasStoragePolicyID() { 5830 return ((bitField0_ & 0x00000100) == 0x00000100); 5831 } 5832 /** 5833 * <code>optional uint32 storagePolicyID = 10;</code> 5834 */ 5835 public int getStoragePolicyID() { 5836 return storagePolicyID_; 5837 } 5838 5839 private void initFields() { 5840 replication_ = 0; 5841 modificationTime_ = 0L; 5842 accessTime_ = 0L; 5843 preferredBlockSize_ = 0L; 5844 permission_ = 0L; 5845 blocks_ = java.util.Collections.emptyList(); 5846 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 5847 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 5848 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 5849 storagePolicyID_ = 0; 5850 } 5851 private byte memoizedIsInitialized = -1; 5852 public final boolean isInitialized() { 5853 byte isInitialized = memoizedIsInitialized; 5854 if (isInitialized != -1) return isInitialized == 1; 5855 5856 for (int i = 0; i < getBlocksCount(); i++) { 5857 if (!getBlocks(i).isInitialized()) { 5858 memoizedIsInitialized = 0; 5859 return false; 5860 } 5861 } 5862 if (hasXAttrs()) { 5863 if (!getXAttrs().isInitialized()) { 5864 memoizedIsInitialized = 0; 5865 return false; 5866 } 5867 } 5868 memoizedIsInitialized = 1; 5869 return true; 5870 } 5871 5872 public void writeTo(com.google.protobuf.CodedOutputStream output) 5873 throws java.io.IOException { 5874 getSerializedSize(); 5875 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5876 output.writeUInt32(1, replication_); 5877 } 5878 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5879 output.writeUInt64(2, modificationTime_); 5880 } 5881 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5882 output.writeUInt64(3, accessTime_); 5883 } 5884 if (((bitField0_ & 0x00000008) == 0x00000008)) { 5885 output.writeUInt64(4, preferredBlockSize_); 5886 } 5887 if (((bitField0_ & 0x00000010) == 0x00000010)) { 5888 output.writeFixed64(5, permission_); 5889 } 5890 for (int i = 0; i < blocks_.size(); i++) { 5891 output.writeMessage(6, blocks_.get(i)); 5892 } 5893 if (((bitField0_ & 0x00000020) == 0x00000020)) { 5894 output.writeMessage(7, fileUC_); 5895 } 5896 if (((bitField0_ & 0x00000040) == 0x00000040)) { 5897 output.writeMessage(8, acl_); 5898 } 5899 if (((bitField0_ & 0x00000080) == 0x00000080)) { 5900 output.writeMessage(9, xAttrs_); 5901 } 5902 if (((bitField0_ & 0x00000100) == 0x00000100)) { 5903 output.writeUInt32(10, storagePolicyID_); 5904 } 5905 getUnknownFields().writeTo(output); 5906 } 5907 5908 private int memoizedSerializedSize = -1; 5909 public int getSerializedSize() { 5910 int size = memoizedSerializedSize; 5911 if (size != -1) return size; 5912 5913 size = 0; 5914 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5915 size += com.google.protobuf.CodedOutputStream 5916 .computeUInt32Size(1, replication_); 5917 } 5918 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5919 size += com.google.protobuf.CodedOutputStream 5920 .computeUInt64Size(2, modificationTime_); 5921 } 5922 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5923 size += com.google.protobuf.CodedOutputStream 5924 .computeUInt64Size(3, accessTime_); 5925 } 5926 if (((bitField0_ & 0x00000008) == 0x00000008)) { 5927 size += com.google.protobuf.CodedOutputStream 5928 .computeUInt64Size(4, preferredBlockSize_); 5929 } 5930 if (((bitField0_ & 0x00000010) == 0x00000010)) { 5931 size += com.google.protobuf.CodedOutputStream 5932 .computeFixed64Size(5, permission_); 5933 } 5934 for (int i = 0; i < blocks_.size(); i++) { 5935 size += com.google.protobuf.CodedOutputStream 5936 .computeMessageSize(6, blocks_.get(i)); 5937 } 5938 if (((bitField0_ & 0x00000020) == 0x00000020)) { 5939 size += com.google.protobuf.CodedOutputStream 5940 .computeMessageSize(7, fileUC_); 5941 } 5942 if (((bitField0_ & 0x00000040) == 0x00000040)) { 5943 size += com.google.protobuf.CodedOutputStream 5944 .computeMessageSize(8, acl_); 5945 } 5946 if (((bitField0_ & 0x00000080) == 0x00000080)) { 5947 size += com.google.protobuf.CodedOutputStream 5948 .computeMessageSize(9, xAttrs_); 5949 } 5950 if (((bitField0_ & 0x00000100) == 0x00000100)) { 5951 size += com.google.protobuf.CodedOutputStream 5952 .computeUInt32Size(10, storagePolicyID_); 5953 } 5954 size += getUnknownFields().getSerializedSize(); 5955 memoizedSerializedSize = size; 5956 return size; 5957 } 5958 5959 private static final long serialVersionUID = 0L; 5960 @java.lang.Override 5961 protected java.lang.Object writeReplace() 5962 throws java.io.ObjectStreamException { 5963 return super.writeReplace(); 5964 } 5965 5966 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 5967 com.google.protobuf.ByteString data) 5968 throws com.google.protobuf.InvalidProtocolBufferException { 5969 return PARSER.parseFrom(data); 5970 } 5971 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 5972 com.google.protobuf.ByteString data, 5973 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5974 throws com.google.protobuf.InvalidProtocolBufferException { 5975 return PARSER.parseFrom(data, extensionRegistry); 5976 } 5977 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom(byte[] data) 5978 throws com.google.protobuf.InvalidProtocolBufferException { 5979 return PARSER.parseFrom(data); 5980 } 5981 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 5982 byte[] data, 5983 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5984 throws com.google.protobuf.InvalidProtocolBufferException { 5985 return PARSER.parseFrom(data, extensionRegistry); 5986 } 5987 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom(java.io.InputStream input) 5988 throws java.io.IOException { 5989 return PARSER.parseFrom(input); 5990 } 5991 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 5992 java.io.InputStream input, 5993 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5994 throws java.io.IOException { 5995 return PARSER.parseFrom(input, extensionRegistry); 5996 } 5997 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseDelimitedFrom(java.io.InputStream input) 5998 throws java.io.IOException { 5999 return PARSER.parseDelimitedFrom(input); 6000 } 6001 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseDelimitedFrom( 6002 java.io.InputStream input, 6003 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6004 throws java.io.IOException { 6005 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6006 } 6007 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 6008 com.google.protobuf.CodedInputStream input) 6009 throws java.io.IOException { 6010 return PARSER.parseFrom(input); 6011 } 6012 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parseFrom( 6013 com.google.protobuf.CodedInputStream input, 6014 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6015 throws java.io.IOException { 6016 return PARSER.parseFrom(input, extensionRegistry); 6017 } 6018 6019 public static Builder newBuilder() { return Builder.create(); } 6020 public Builder newBuilderForType() { return newBuilder(); } 6021 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile prototype) { 6022 return newBuilder().mergeFrom(prototype); 6023 } 6024 public Builder toBuilder() { return newBuilder(this); } 6025 6026 @java.lang.Override 6027 protected Builder newBuilderForType( 6028 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6029 Builder builder = new Builder(parent); 6030 return builder; 6031 } 6032 /** 6033 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeFile} 6034 */ 6035 public static final class Builder extends 6036 com.google.protobuf.GeneratedMessage.Builder<Builder> 6037 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder { 6038 public static final com.google.protobuf.Descriptors.Descriptor 6039 getDescriptor() { 6040 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 6041 } 6042 6043 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 6044 internalGetFieldAccessorTable() { 6045 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable 6046 .ensureFieldAccessorsInitialized( 6047 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder.class); 6048 } 6049 6050 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder() 6051 private Builder() { 6052 maybeForceBuilderInitialization(); 6053 } 6054 6055 private Builder( 6056 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6057 super(parent); 6058 maybeForceBuilderInitialization(); 6059 } 6060 private void maybeForceBuilderInitialization() { 6061 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6062 getBlocksFieldBuilder(); 6063 getFileUCFieldBuilder(); 6064 getAclFieldBuilder(); 6065 getXAttrsFieldBuilder(); 6066 } 6067 } 6068 private static Builder create() { 6069 return new Builder(); 6070 } 6071 6072 public Builder clear() { 6073 super.clear(); 6074 replication_ = 0; 6075 bitField0_ = (bitField0_ & ~0x00000001); 6076 modificationTime_ = 0L; 6077 bitField0_ = (bitField0_ & ~0x00000002); 6078 accessTime_ = 0L; 6079 bitField0_ = (bitField0_ & ~0x00000004); 6080 preferredBlockSize_ = 0L; 6081 bitField0_ = (bitField0_ & ~0x00000008); 6082 permission_ = 0L; 6083 bitField0_ = (bitField0_ & ~0x00000010); 6084 if (blocksBuilder_ == null) { 6085 blocks_ = java.util.Collections.emptyList(); 6086 bitField0_ = (bitField0_ & ~0x00000020); 6087 } else { 6088 blocksBuilder_.clear(); 6089 } 6090 if (fileUCBuilder_ == null) { 6091 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 6092 } else { 6093 fileUCBuilder_.clear(); 6094 } 6095 bitField0_ = (bitField0_ & ~0x00000040); 6096 if (aclBuilder_ == null) { 6097 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6098 } else { 6099 aclBuilder_.clear(); 6100 } 6101 bitField0_ = (bitField0_ & ~0x00000080); 6102 if (xAttrsBuilder_ == null) { 6103 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 6104 } else { 6105 xAttrsBuilder_.clear(); 6106 } 6107 bitField0_ = (bitField0_ & ~0x00000100); 6108 storagePolicyID_ = 0; 6109 bitField0_ = (bitField0_ & ~0x00000200); 6110 return this; 6111 } 6112 6113 public Builder clone() { 6114 return create().mergeFrom(buildPartial()); 6115 } 6116 6117 public com.google.protobuf.Descriptors.Descriptor 6118 getDescriptorForType() { 6119 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 6120 } 6121 6122 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getDefaultInstanceForType() { 6123 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 6124 } 6125 6126 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile build() { 6127 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile result = buildPartial(); 6128 if (!result.isInitialized()) { 6129 throw newUninitializedMessageException(result); 6130 } 6131 return result; 6132 } 6133 6134 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile buildPartial() { 6135 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile(this); 6136 int from_bitField0_ = bitField0_; 6137 int to_bitField0_ = 0; 6138 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 6139 to_bitField0_ |= 0x00000001; 6140 } 6141 result.replication_ = replication_; 6142 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 6143 to_bitField0_ |= 0x00000002; 6144 } 6145 result.modificationTime_ = modificationTime_; 6146 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 6147 to_bitField0_ |= 0x00000004; 6148 } 6149 result.accessTime_ = accessTime_; 6150 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 6151 to_bitField0_ |= 0x00000008; 6152 } 6153 result.preferredBlockSize_ = preferredBlockSize_; 6154 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 6155 to_bitField0_ |= 0x00000010; 6156 } 6157 result.permission_ = permission_; 6158 if (blocksBuilder_ == null) { 6159 if (((bitField0_ & 0x00000020) == 0x00000020)) { 6160 blocks_ = java.util.Collections.unmodifiableList(blocks_); 6161 bitField0_ = (bitField0_ & ~0x00000020); 6162 } 6163 result.blocks_ = blocks_; 6164 } else { 6165 result.blocks_ = blocksBuilder_.build(); 6166 } 6167 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 6168 to_bitField0_ |= 0x00000020; 6169 } 6170 if (fileUCBuilder_ == null) { 6171 result.fileUC_ = fileUC_; 6172 } else { 6173 result.fileUC_ = fileUCBuilder_.build(); 6174 } 6175 if (((from_bitField0_ & 0x00000080) == 0x00000080)) { 6176 to_bitField0_ |= 0x00000040; 6177 } 6178 if (aclBuilder_ == null) { 6179 result.acl_ = acl_; 6180 } else { 6181 result.acl_ = aclBuilder_.build(); 6182 } 6183 if (((from_bitField0_ & 0x00000100) == 0x00000100)) { 6184 to_bitField0_ |= 0x00000080; 6185 } 6186 if (xAttrsBuilder_ == null) { 6187 result.xAttrs_ = xAttrs_; 6188 } else { 6189 result.xAttrs_ = xAttrsBuilder_.build(); 6190 } 6191 if (((from_bitField0_ & 0x00000200) == 0x00000200)) { 6192 to_bitField0_ |= 0x00000100; 6193 } 6194 result.storagePolicyID_ = storagePolicyID_; 6195 result.bitField0_ = to_bitField0_; 6196 onBuilt(); 6197 return result; 6198 } 6199 6200 public Builder mergeFrom(com.google.protobuf.Message other) { 6201 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile) { 6202 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile)other); 6203 } else { 6204 super.mergeFrom(other); 6205 return this; 6206 } 6207 } 6208 6209 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile other) { 6210 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) return this; 6211 if (other.hasReplication()) { 6212 setReplication(other.getReplication()); 6213 } 6214 if (other.hasModificationTime()) { 6215 setModificationTime(other.getModificationTime()); 6216 } 6217 if (other.hasAccessTime()) { 6218 setAccessTime(other.getAccessTime()); 6219 } 6220 if (other.hasPreferredBlockSize()) { 6221 setPreferredBlockSize(other.getPreferredBlockSize()); 6222 } 6223 if (other.hasPermission()) { 6224 setPermission(other.getPermission()); 6225 } 6226 if (blocksBuilder_ == null) { 6227 if (!other.blocks_.isEmpty()) { 6228 if (blocks_.isEmpty()) { 6229 blocks_ = other.blocks_; 6230 bitField0_ = (bitField0_ & ~0x00000020); 6231 } else { 6232 ensureBlocksIsMutable(); 6233 blocks_.addAll(other.blocks_); 6234 } 6235 onChanged(); 6236 } 6237 } else { 6238 if (!other.blocks_.isEmpty()) { 6239 if (blocksBuilder_.isEmpty()) { 6240 blocksBuilder_.dispose(); 6241 blocksBuilder_ = null; 6242 blocks_ = other.blocks_; 6243 bitField0_ = (bitField0_ & ~0x00000020); 6244 blocksBuilder_ = 6245 com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? 6246 getBlocksFieldBuilder() : null; 6247 } else { 6248 blocksBuilder_.addAllMessages(other.blocks_); 6249 } 6250 } 6251 } 6252 if (other.hasFileUC()) { 6253 mergeFileUC(other.getFileUC()); 6254 } 6255 if (other.hasAcl()) { 6256 mergeAcl(other.getAcl()); 6257 } 6258 if (other.hasXAttrs()) { 6259 mergeXAttrs(other.getXAttrs()); 6260 } 6261 if (other.hasStoragePolicyID()) { 6262 setStoragePolicyID(other.getStoragePolicyID()); 6263 } 6264 this.mergeUnknownFields(other.getUnknownFields()); 6265 return this; 6266 } 6267 6268 public final boolean isInitialized() { 6269 for (int i = 0; i < getBlocksCount(); i++) { 6270 if (!getBlocks(i).isInitialized()) { 6271 6272 return false; 6273 } 6274 } 6275 if (hasXAttrs()) { 6276 if (!getXAttrs().isInitialized()) { 6277 6278 return false; 6279 } 6280 } 6281 return true; 6282 } 6283 6284 public Builder mergeFrom( 6285 com.google.protobuf.CodedInputStream input, 6286 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6287 throws java.io.IOException { 6288 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile parsedMessage = null; 6289 try { 6290 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6291 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6292 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile) e.getUnfinishedMessage(); 6293 throw e; 6294 } finally { 6295 if (parsedMessage != null) { 6296 mergeFrom(parsedMessage); 6297 } 6298 } 6299 return this; 6300 } 6301 private int bitField0_; 6302 6303 // optional uint32 replication = 1; 6304 private int replication_ ; 6305 /** 6306 * <code>optional uint32 replication = 1;</code> 6307 */ 6308 public boolean hasReplication() { 6309 return ((bitField0_ & 0x00000001) == 0x00000001); 6310 } 6311 /** 6312 * <code>optional uint32 replication = 1;</code> 6313 */ 6314 public int getReplication() { 6315 return replication_; 6316 } 6317 /** 6318 * <code>optional uint32 replication = 1;</code> 6319 */ 6320 public Builder setReplication(int value) { 6321 bitField0_ |= 0x00000001; 6322 replication_ = value; 6323 onChanged(); 6324 return this; 6325 } 6326 /** 6327 * <code>optional uint32 replication = 1;</code> 6328 */ 6329 public Builder clearReplication() { 6330 bitField0_ = (bitField0_ & ~0x00000001); 6331 replication_ = 0; 6332 onChanged(); 6333 return this; 6334 } 6335 6336 // optional uint64 modificationTime = 2; 6337 private long modificationTime_ ; 6338 /** 6339 * <code>optional uint64 modificationTime = 2;</code> 6340 */ 6341 public boolean hasModificationTime() { 6342 return ((bitField0_ & 0x00000002) == 0x00000002); 6343 } 6344 /** 6345 * <code>optional uint64 modificationTime = 2;</code> 6346 */ 6347 public long getModificationTime() { 6348 return modificationTime_; 6349 } 6350 /** 6351 * <code>optional uint64 modificationTime = 2;</code> 6352 */ 6353 public Builder setModificationTime(long value) { 6354 bitField0_ |= 0x00000002; 6355 modificationTime_ = value; 6356 onChanged(); 6357 return this; 6358 } 6359 /** 6360 * <code>optional uint64 modificationTime = 2;</code> 6361 */ 6362 public Builder clearModificationTime() { 6363 bitField0_ = (bitField0_ & ~0x00000002); 6364 modificationTime_ = 0L; 6365 onChanged(); 6366 return this; 6367 } 6368 6369 // optional uint64 accessTime = 3; 6370 private long accessTime_ ; 6371 /** 6372 * <code>optional uint64 accessTime = 3;</code> 6373 */ 6374 public boolean hasAccessTime() { 6375 return ((bitField0_ & 0x00000004) == 0x00000004); 6376 } 6377 /** 6378 * <code>optional uint64 accessTime = 3;</code> 6379 */ 6380 public long getAccessTime() { 6381 return accessTime_; 6382 } 6383 /** 6384 * <code>optional uint64 accessTime = 3;</code> 6385 */ 6386 public Builder setAccessTime(long value) { 6387 bitField0_ |= 0x00000004; 6388 accessTime_ = value; 6389 onChanged(); 6390 return this; 6391 } 6392 /** 6393 * <code>optional uint64 accessTime = 3;</code> 6394 */ 6395 public Builder clearAccessTime() { 6396 bitField0_ = (bitField0_ & ~0x00000004); 6397 accessTime_ = 0L; 6398 onChanged(); 6399 return this; 6400 } 6401 6402 // optional uint64 preferredBlockSize = 4; 6403 private long preferredBlockSize_ ; 6404 /** 6405 * <code>optional uint64 preferredBlockSize = 4;</code> 6406 */ 6407 public boolean hasPreferredBlockSize() { 6408 return ((bitField0_ & 0x00000008) == 0x00000008); 6409 } 6410 /** 6411 * <code>optional uint64 preferredBlockSize = 4;</code> 6412 */ 6413 public long getPreferredBlockSize() { 6414 return preferredBlockSize_; 6415 } 6416 /** 6417 * <code>optional uint64 preferredBlockSize = 4;</code> 6418 */ 6419 public Builder setPreferredBlockSize(long value) { 6420 bitField0_ |= 0x00000008; 6421 preferredBlockSize_ = value; 6422 onChanged(); 6423 return this; 6424 } 6425 /** 6426 * <code>optional uint64 preferredBlockSize = 4;</code> 6427 */ 6428 public Builder clearPreferredBlockSize() { 6429 bitField0_ = (bitField0_ & ~0x00000008); 6430 preferredBlockSize_ = 0L; 6431 onChanged(); 6432 return this; 6433 } 6434 6435 // optional fixed64 permission = 5; 6436 private long permission_ ; 6437 /** 6438 * <code>optional fixed64 permission = 5;</code> 6439 */ 6440 public boolean hasPermission() { 6441 return ((bitField0_ & 0x00000010) == 0x00000010); 6442 } 6443 /** 6444 * <code>optional fixed64 permission = 5;</code> 6445 */ 6446 public long getPermission() { 6447 return permission_; 6448 } 6449 /** 6450 * <code>optional fixed64 permission = 5;</code> 6451 */ 6452 public Builder setPermission(long value) { 6453 bitField0_ |= 0x00000010; 6454 permission_ = value; 6455 onChanged(); 6456 return this; 6457 } 6458 /** 6459 * <code>optional fixed64 permission = 5;</code> 6460 */ 6461 public Builder clearPermission() { 6462 bitField0_ = (bitField0_ & ~0x00000010); 6463 permission_ = 0L; 6464 onChanged(); 6465 return this; 6466 } 6467 6468 // repeated .hadoop.hdfs.BlockProto blocks = 6; 6469 private java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> blocks_ = 6470 java.util.Collections.emptyList(); 6471 private void ensureBlocksIsMutable() { 6472 if (!((bitField0_ & 0x00000020) == 0x00000020)) { 6473 blocks_ = new java.util.ArrayList<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto>(blocks_); 6474 bitField0_ |= 0x00000020; 6475 } 6476 } 6477 6478 private com.google.protobuf.RepeatedFieldBuilder< 6479 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> blocksBuilder_; 6480 6481 /** 6482 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6483 */ 6484 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> getBlocksList() { 6485 if (blocksBuilder_ == null) { 6486 return java.util.Collections.unmodifiableList(blocks_); 6487 } else { 6488 return blocksBuilder_.getMessageList(); 6489 } 6490 } 6491 /** 6492 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6493 */ 6494 public int getBlocksCount() { 6495 if (blocksBuilder_ == null) { 6496 return blocks_.size(); 6497 } else { 6498 return blocksBuilder_.getCount(); 6499 } 6500 } 6501 /** 6502 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6503 */ 6504 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { 6505 if (blocksBuilder_ == null) { 6506 return blocks_.get(index); 6507 } else { 6508 return blocksBuilder_.getMessage(index); 6509 } 6510 } 6511 /** 6512 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6513 */ 6514 public Builder setBlocks( 6515 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 6516 if (blocksBuilder_ == null) { 6517 if (value == null) { 6518 throw new NullPointerException(); 6519 } 6520 ensureBlocksIsMutable(); 6521 blocks_.set(index, value); 6522 onChanged(); 6523 } else { 6524 blocksBuilder_.setMessage(index, value); 6525 } 6526 return this; 6527 } 6528 /** 6529 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6530 */ 6531 public Builder setBlocks( 6532 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 6533 if (blocksBuilder_ == null) { 6534 ensureBlocksIsMutable(); 6535 blocks_.set(index, builderForValue.build()); 6536 onChanged(); 6537 } else { 6538 blocksBuilder_.setMessage(index, builderForValue.build()); 6539 } 6540 return this; 6541 } 6542 /** 6543 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6544 */ 6545 public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 6546 if (blocksBuilder_ == null) { 6547 if (value == null) { 6548 throw new NullPointerException(); 6549 } 6550 ensureBlocksIsMutable(); 6551 blocks_.add(value); 6552 onChanged(); 6553 } else { 6554 blocksBuilder_.addMessage(value); 6555 } 6556 return this; 6557 } 6558 /** 6559 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6560 */ 6561 public Builder addBlocks( 6562 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { 6563 if (blocksBuilder_ == null) { 6564 if (value == null) { 6565 throw new NullPointerException(); 6566 } 6567 ensureBlocksIsMutable(); 6568 blocks_.add(index, value); 6569 onChanged(); 6570 } else { 6571 blocksBuilder_.addMessage(index, value); 6572 } 6573 return this; 6574 } 6575 /** 6576 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6577 */ 6578 public Builder addBlocks( 6579 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 6580 if (blocksBuilder_ == null) { 6581 ensureBlocksIsMutable(); 6582 blocks_.add(builderForValue.build()); 6583 onChanged(); 6584 } else { 6585 blocksBuilder_.addMessage(builderForValue.build()); 6586 } 6587 return this; 6588 } 6589 /** 6590 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6591 */ 6592 public Builder addBlocks( 6593 int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { 6594 if (blocksBuilder_ == null) { 6595 ensureBlocksIsMutable(); 6596 blocks_.add(index, builderForValue.build()); 6597 onChanged(); 6598 } else { 6599 blocksBuilder_.addMessage(index, builderForValue.build()); 6600 } 6601 return this; 6602 } 6603 /** 6604 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6605 */ 6606 public Builder addAllBlocks( 6607 java.lang.Iterable<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto> values) { 6608 if (blocksBuilder_ == null) { 6609 ensureBlocksIsMutable(); 6610 super.addAll(values, blocks_); 6611 onChanged(); 6612 } else { 6613 blocksBuilder_.addAllMessages(values); 6614 } 6615 return this; 6616 } 6617 /** 6618 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6619 */ 6620 public Builder clearBlocks() { 6621 if (blocksBuilder_ == null) { 6622 blocks_ = java.util.Collections.emptyList(); 6623 bitField0_ = (bitField0_ & ~0x00000020); 6624 onChanged(); 6625 } else { 6626 blocksBuilder_.clear(); 6627 } 6628 return this; 6629 } 6630 /** 6631 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6632 */ 6633 public Builder removeBlocks(int index) { 6634 if (blocksBuilder_ == null) { 6635 ensureBlocksIsMutable(); 6636 blocks_.remove(index); 6637 onChanged(); 6638 } else { 6639 blocksBuilder_.remove(index); 6640 } 6641 return this; 6642 } 6643 /** 6644 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6645 */ 6646 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder getBlocksBuilder( 6647 int index) { 6648 return getBlocksFieldBuilder().getBuilder(index); 6649 } 6650 /** 6651 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6652 */ 6653 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( 6654 int index) { 6655 if (blocksBuilder_ == null) { 6656 return blocks_.get(index); } else { 6657 return blocksBuilder_.getMessageOrBuilder(index); 6658 } 6659 } 6660 /** 6661 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6662 */ 6663 public java.util.List<? extends org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 6664 getBlocksOrBuilderList() { 6665 if (blocksBuilder_ != null) { 6666 return blocksBuilder_.getMessageOrBuilderList(); 6667 } else { 6668 return java.util.Collections.unmodifiableList(blocks_); 6669 } 6670 } 6671 /** 6672 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6673 */ 6674 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { 6675 return getBlocksFieldBuilder().addBuilder( 6676 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); 6677 } 6678 /** 6679 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6680 */ 6681 public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( 6682 int index) { 6683 return getBlocksFieldBuilder().addBuilder( 6684 index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); 6685 } 6686 /** 6687 * <code>repeated .hadoop.hdfs.BlockProto blocks = 6;</code> 6688 */ 6689 public java.util.List<org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder> 6690 getBlocksBuilderList() { 6691 return getBlocksFieldBuilder().getBuilderList(); 6692 } 6693 private com.google.protobuf.RepeatedFieldBuilder< 6694 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> 6695 getBlocksFieldBuilder() { 6696 if (blocksBuilder_ == null) { 6697 blocksBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< 6698 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder>( 6699 blocks_, 6700 ((bitField0_ & 0x00000020) == 0x00000020), 6701 getParentForChildren(), 6702 isClean()); 6703 blocks_ = null; 6704 } 6705 return blocksBuilder_; 6706 } 6707 6708 // optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7; 6709 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 6710 private com.google.protobuf.SingleFieldBuilder< 6711 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder> fileUCBuilder_; 6712 /** 6713 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6714 */ 6715 public boolean hasFileUC() { 6716 return ((bitField0_ & 0x00000040) == 0x00000040); 6717 } 6718 /** 6719 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6720 */ 6721 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature getFileUC() { 6722 if (fileUCBuilder_ == null) { 6723 return fileUC_; 6724 } else { 6725 return fileUCBuilder_.getMessage(); 6726 } 6727 } 6728 /** 6729 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6730 */ 6731 public Builder setFileUC(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature value) { 6732 if (fileUCBuilder_ == null) { 6733 if (value == null) { 6734 throw new NullPointerException(); 6735 } 6736 fileUC_ = value; 6737 onChanged(); 6738 } else { 6739 fileUCBuilder_.setMessage(value); 6740 } 6741 bitField0_ |= 0x00000040; 6742 return this; 6743 } 6744 /** 6745 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6746 */ 6747 public Builder setFileUC( 6748 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder builderForValue) { 6749 if (fileUCBuilder_ == null) { 6750 fileUC_ = builderForValue.build(); 6751 onChanged(); 6752 } else { 6753 fileUCBuilder_.setMessage(builderForValue.build()); 6754 } 6755 bitField0_ |= 0x00000040; 6756 return this; 6757 } 6758 /** 6759 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6760 */ 6761 public Builder mergeFileUC(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature value) { 6762 if (fileUCBuilder_ == null) { 6763 if (((bitField0_ & 0x00000040) == 0x00000040) && 6764 fileUC_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance()) { 6765 fileUC_ = 6766 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.newBuilder(fileUC_).mergeFrom(value).buildPartial(); 6767 } else { 6768 fileUC_ = value; 6769 } 6770 onChanged(); 6771 } else { 6772 fileUCBuilder_.mergeFrom(value); 6773 } 6774 bitField0_ |= 0x00000040; 6775 return this; 6776 } 6777 /** 6778 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6779 */ 6780 public Builder clearFileUC() { 6781 if (fileUCBuilder_ == null) { 6782 fileUC_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.getDefaultInstance(); 6783 onChanged(); 6784 } else { 6785 fileUCBuilder_.clear(); 6786 } 6787 bitField0_ = (bitField0_ & ~0x00000040); 6788 return this; 6789 } 6790 /** 6791 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6792 */ 6793 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder getFileUCBuilder() { 6794 bitField0_ |= 0x00000040; 6795 onChanged(); 6796 return getFileUCFieldBuilder().getBuilder(); 6797 } 6798 /** 6799 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6800 */ 6801 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder getFileUCOrBuilder() { 6802 if (fileUCBuilder_ != null) { 6803 return fileUCBuilder_.getMessageOrBuilder(); 6804 } else { 6805 return fileUC_; 6806 } 6807 } 6808 /** 6809 * <code>optional .hadoop.hdfs.fsimage.INodeSection.FileUnderConstructionFeature fileUC = 7;</code> 6810 */ 6811 private com.google.protobuf.SingleFieldBuilder< 6812 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder> 6813 getFileUCFieldBuilder() { 6814 if (fileUCBuilder_ == null) { 6815 fileUCBuilder_ = new com.google.protobuf.SingleFieldBuilder< 6816 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeature.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.FileUnderConstructionFeatureOrBuilder>( 6817 fileUC_, 6818 getParentForChildren(), 6819 isClean()); 6820 fileUC_ = null; 6821 } 6822 return fileUCBuilder_; 6823 } 6824 6825 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8; 6826 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6827 private com.google.protobuf.SingleFieldBuilder< 6828 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> aclBuilder_; 6829 /** 6830 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6831 */ 6832 public boolean hasAcl() { 6833 return ((bitField0_ & 0x00000080) == 0x00000080); 6834 } 6835 /** 6836 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6837 */ 6838 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 6839 if (aclBuilder_ == null) { 6840 return acl_; 6841 } else { 6842 return aclBuilder_.getMessage(); 6843 } 6844 } 6845 /** 6846 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6847 */ 6848 public Builder setAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 6849 if (aclBuilder_ == null) { 6850 if (value == null) { 6851 throw new NullPointerException(); 6852 } 6853 acl_ = value; 6854 onChanged(); 6855 } else { 6856 aclBuilder_.setMessage(value); 6857 } 6858 bitField0_ |= 0x00000080; 6859 return this; 6860 } 6861 /** 6862 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6863 */ 6864 public Builder setAcl( 6865 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder builderForValue) { 6866 if (aclBuilder_ == null) { 6867 acl_ = builderForValue.build(); 6868 onChanged(); 6869 } else { 6870 aclBuilder_.setMessage(builderForValue.build()); 6871 } 6872 bitField0_ |= 0x00000080; 6873 return this; 6874 } 6875 /** 6876 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6877 */ 6878 public Builder mergeAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 6879 if (aclBuilder_ == null) { 6880 if (((bitField0_ & 0x00000080) == 0x00000080) && 6881 acl_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) { 6882 acl_ = 6883 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder(acl_).mergeFrom(value).buildPartial(); 6884 } else { 6885 acl_ = value; 6886 } 6887 onChanged(); 6888 } else { 6889 aclBuilder_.mergeFrom(value); 6890 } 6891 bitField0_ |= 0x00000080; 6892 return this; 6893 } 6894 /** 6895 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6896 */ 6897 public Builder clearAcl() { 6898 if (aclBuilder_ == null) { 6899 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 6900 onChanged(); 6901 } else { 6902 aclBuilder_.clear(); 6903 } 6904 bitField0_ = (bitField0_ & ~0x00000080); 6905 return this; 6906 } 6907 /** 6908 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6909 */ 6910 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder getAclBuilder() { 6911 bitField0_ |= 0x00000080; 6912 onChanged(); 6913 return getAclFieldBuilder().getBuilder(); 6914 } 6915 /** 6916 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6917 */ 6918 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 6919 if (aclBuilder_ != null) { 6920 return aclBuilder_.getMessageOrBuilder(); 6921 } else { 6922 return acl_; 6923 } 6924 } 6925 /** 6926 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 8;</code> 6927 */ 6928 private com.google.protobuf.SingleFieldBuilder< 6929 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> 6930 getAclFieldBuilder() { 6931 if (aclBuilder_ == null) { 6932 aclBuilder_ = new com.google.protobuf.SingleFieldBuilder< 6933 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder>( 6934 acl_, 6935 getParentForChildren(), 6936 isClean()); 6937 acl_ = null; 6938 } 6939 return aclBuilder_; 6940 } 6941 6942 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9; 6943 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 6944 private com.google.protobuf.SingleFieldBuilder< 6945 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder> xAttrsBuilder_; 6946 /** 6947 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 6948 */ 6949 public boolean hasXAttrs() { 6950 return ((bitField0_ & 0x00000100) == 0x00000100); 6951 } 6952 /** 6953 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 6954 */ 6955 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs() { 6956 if (xAttrsBuilder_ == null) { 6957 return xAttrs_; 6958 } else { 6959 return xAttrsBuilder_.getMessage(); 6960 } 6961 } 6962 /** 6963 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 6964 */ 6965 public Builder setXAttrs(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto value) { 6966 if (xAttrsBuilder_ == null) { 6967 if (value == null) { 6968 throw new NullPointerException(); 6969 } 6970 xAttrs_ = value; 6971 onChanged(); 6972 } else { 6973 xAttrsBuilder_.setMessage(value); 6974 } 6975 bitField0_ |= 0x00000100; 6976 return this; 6977 } 6978 /** 6979 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 6980 */ 6981 public Builder setXAttrs( 6982 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder builderForValue) { 6983 if (xAttrsBuilder_ == null) { 6984 xAttrs_ = builderForValue.build(); 6985 onChanged(); 6986 } else { 6987 xAttrsBuilder_.setMessage(builderForValue.build()); 6988 } 6989 bitField0_ |= 0x00000100; 6990 return this; 6991 } 6992 /** 6993 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 6994 */ 6995 public Builder mergeXAttrs(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto value) { 6996 if (xAttrsBuilder_ == null) { 6997 if (((bitField0_ & 0x00000100) == 0x00000100) && 6998 xAttrs_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance()) { 6999 xAttrs_ = 7000 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.newBuilder(xAttrs_).mergeFrom(value).buildPartial(); 7001 } else { 7002 xAttrs_ = value; 7003 } 7004 onChanged(); 7005 } else { 7006 xAttrsBuilder_.mergeFrom(value); 7007 } 7008 bitField0_ |= 0x00000100; 7009 return this; 7010 } 7011 /** 7012 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 7013 */ 7014 public Builder clearXAttrs() { 7015 if (xAttrsBuilder_ == null) { 7016 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 7017 onChanged(); 7018 } else { 7019 xAttrsBuilder_.clear(); 7020 } 7021 bitField0_ = (bitField0_ & ~0x00000100); 7022 return this; 7023 } 7024 /** 7025 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 7026 */ 7027 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder getXAttrsBuilder() { 7028 bitField0_ |= 0x00000100; 7029 onChanged(); 7030 return getXAttrsFieldBuilder().getBuilder(); 7031 } 7032 /** 7033 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 7034 */ 7035 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder() { 7036 if (xAttrsBuilder_ != null) { 7037 return xAttrsBuilder_.getMessageOrBuilder(); 7038 } else { 7039 return xAttrs_; 7040 } 7041 } 7042 /** 7043 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 9;</code> 7044 */ 7045 private com.google.protobuf.SingleFieldBuilder< 7046 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder> 7047 getXAttrsFieldBuilder() { 7048 if (xAttrsBuilder_ == null) { 7049 xAttrsBuilder_ = new com.google.protobuf.SingleFieldBuilder< 7050 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder>( 7051 xAttrs_, 7052 getParentForChildren(), 7053 isClean()); 7054 xAttrs_ = null; 7055 } 7056 return xAttrsBuilder_; 7057 } 7058 7059 // optional uint32 storagePolicyID = 10; 7060 private int storagePolicyID_ ; 7061 /** 7062 * <code>optional uint32 storagePolicyID = 10;</code> 7063 */ 7064 public boolean hasStoragePolicyID() { 7065 return ((bitField0_ & 0x00000200) == 0x00000200); 7066 } 7067 /** 7068 * <code>optional uint32 storagePolicyID = 10;</code> 7069 */ 7070 public int getStoragePolicyID() { 7071 return storagePolicyID_; 7072 } 7073 /** 7074 * <code>optional uint32 storagePolicyID = 10;</code> 7075 */ 7076 public Builder setStoragePolicyID(int value) { 7077 bitField0_ |= 0x00000200; 7078 storagePolicyID_ = value; 7079 onChanged(); 7080 return this; 7081 } 7082 /** 7083 * <code>optional uint32 storagePolicyID = 10;</code> 7084 */ 7085 public Builder clearStoragePolicyID() { 7086 bitField0_ = (bitField0_ & ~0x00000200); 7087 storagePolicyID_ = 0; 7088 onChanged(); 7089 return this; 7090 } 7091 7092 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeFile) 7093 } 7094 7095 static { 7096 defaultInstance = new INodeFile(true); 7097 defaultInstance.initFields(); 7098 } 7099 7100 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeFile) 7101 } 7102 7103 public interface INodeDirectoryOrBuilder 7104 extends com.google.protobuf.MessageOrBuilder { 7105 7106 // optional uint64 modificationTime = 1; 7107 /** 7108 * <code>optional uint64 modificationTime = 1;</code> 7109 */ 7110 boolean hasModificationTime(); 7111 /** 7112 * <code>optional uint64 modificationTime = 1;</code> 7113 */ 7114 long getModificationTime(); 7115 7116 // optional uint64 nsQuota = 2; 7117 /** 7118 * <code>optional uint64 nsQuota = 2;</code> 7119 * 7120 * <pre> 7121 * namespace quota 7122 * </pre> 7123 */ 7124 boolean hasNsQuota(); 7125 /** 7126 * <code>optional uint64 nsQuota = 2;</code> 7127 * 7128 * <pre> 7129 * namespace quota 7130 * </pre> 7131 */ 7132 long getNsQuota(); 7133 7134 // optional uint64 dsQuota = 3; 7135 /** 7136 * <code>optional uint64 dsQuota = 3;</code> 7137 * 7138 * <pre> 7139 * diskspace quota 7140 * </pre> 7141 */ 7142 boolean hasDsQuota(); 7143 /** 7144 * <code>optional uint64 dsQuota = 3;</code> 7145 * 7146 * <pre> 7147 * diskspace quota 7148 * </pre> 7149 */ 7150 long getDsQuota(); 7151 7152 // optional fixed64 permission = 4; 7153 /** 7154 * <code>optional fixed64 permission = 4;</code> 7155 */ 7156 boolean hasPermission(); 7157 /** 7158 * <code>optional fixed64 permission = 4;</code> 7159 */ 7160 long getPermission(); 7161 7162 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 7163 /** 7164 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7165 */ 7166 boolean hasAcl(); 7167 /** 7168 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7169 */ 7170 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl(); 7171 /** 7172 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7173 */ 7174 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder(); 7175 7176 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6; 7177 /** 7178 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7179 */ 7180 boolean hasXAttrs(); 7181 /** 7182 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7183 */ 7184 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs(); 7185 /** 7186 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7187 */ 7188 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder(); 7189 } 7190 /** 7191 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeDirectory} 7192 */ 7193 public static final class INodeDirectory extends 7194 com.google.protobuf.GeneratedMessage 7195 implements INodeDirectoryOrBuilder { 7196 // Use INodeDirectory.newBuilder() to construct. 7197 private INodeDirectory(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7198 super(builder); 7199 this.unknownFields = builder.getUnknownFields(); 7200 } 7201 private INodeDirectory(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7202 7203 private static final INodeDirectory defaultInstance; 7204 public static INodeDirectory getDefaultInstance() { 7205 return defaultInstance; 7206 } 7207 7208 public INodeDirectory getDefaultInstanceForType() { 7209 return defaultInstance; 7210 } 7211 7212 private final com.google.protobuf.UnknownFieldSet unknownFields; 7213 @java.lang.Override 7214 public final com.google.protobuf.UnknownFieldSet 7215 getUnknownFields() { 7216 return this.unknownFields; 7217 } 7218 private INodeDirectory( 7219 com.google.protobuf.CodedInputStream input, 7220 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7221 throws com.google.protobuf.InvalidProtocolBufferException { 7222 initFields(); 7223 int mutable_bitField0_ = 0; 7224 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7225 com.google.protobuf.UnknownFieldSet.newBuilder(); 7226 try { 7227 boolean done = false; 7228 while (!done) { 7229 int tag = input.readTag(); 7230 switch (tag) { 7231 case 0: 7232 done = true; 7233 break; 7234 default: { 7235 if (!parseUnknownField(input, unknownFields, 7236 extensionRegistry, tag)) { 7237 done = true; 7238 } 7239 break; 7240 } 7241 case 8: { 7242 bitField0_ |= 0x00000001; 7243 modificationTime_ = input.readUInt64(); 7244 break; 7245 } 7246 case 16: { 7247 bitField0_ |= 0x00000002; 7248 nsQuota_ = input.readUInt64(); 7249 break; 7250 } 7251 case 24: { 7252 bitField0_ |= 0x00000004; 7253 dsQuota_ = input.readUInt64(); 7254 break; 7255 } 7256 case 33: { 7257 bitField0_ |= 0x00000008; 7258 permission_ = input.readFixed64(); 7259 break; 7260 } 7261 case 42: { 7262 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder subBuilder = null; 7263 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7264 subBuilder = acl_.toBuilder(); 7265 } 7266 acl_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.PARSER, extensionRegistry); 7267 if (subBuilder != null) { 7268 subBuilder.mergeFrom(acl_); 7269 acl_ = subBuilder.buildPartial(); 7270 } 7271 bitField0_ |= 0x00000010; 7272 break; 7273 } 7274 case 50: { 7275 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder subBuilder = null; 7276 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7277 subBuilder = xAttrs_.toBuilder(); 7278 } 7279 xAttrs_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.PARSER, extensionRegistry); 7280 if (subBuilder != null) { 7281 subBuilder.mergeFrom(xAttrs_); 7282 xAttrs_ = subBuilder.buildPartial(); 7283 } 7284 bitField0_ |= 0x00000020; 7285 break; 7286 } 7287 } 7288 } 7289 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7290 throw e.setUnfinishedMessage(this); 7291 } catch (java.io.IOException e) { 7292 throw new com.google.protobuf.InvalidProtocolBufferException( 7293 e.getMessage()).setUnfinishedMessage(this); 7294 } finally { 7295 this.unknownFields = unknownFields.build(); 7296 makeExtensionsImmutable(); 7297 } 7298 } 7299 public static final com.google.protobuf.Descriptors.Descriptor 7300 getDescriptor() { 7301 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 7302 } 7303 7304 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7305 internalGetFieldAccessorTable() { 7306 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable 7307 .ensureFieldAccessorsInitialized( 7308 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder.class); 7309 } 7310 7311 public static com.google.protobuf.Parser<INodeDirectory> PARSER = 7312 new com.google.protobuf.AbstractParser<INodeDirectory>() { 7313 public INodeDirectory parsePartialFrom( 7314 com.google.protobuf.CodedInputStream input, 7315 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7316 throws com.google.protobuf.InvalidProtocolBufferException { 7317 return new INodeDirectory(input, extensionRegistry); 7318 } 7319 }; 7320 7321 @java.lang.Override 7322 public com.google.protobuf.Parser<INodeDirectory> getParserForType() { 7323 return PARSER; 7324 } 7325 7326 private int bitField0_; 7327 // optional uint64 modificationTime = 1; 7328 public static final int MODIFICATIONTIME_FIELD_NUMBER = 1; 7329 private long modificationTime_; 7330 /** 7331 * <code>optional uint64 modificationTime = 1;</code> 7332 */ 7333 public boolean hasModificationTime() { 7334 return ((bitField0_ & 0x00000001) == 0x00000001); 7335 } 7336 /** 7337 * <code>optional uint64 modificationTime = 1;</code> 7338 */ 7339 public long getModificationTime() { 7340 return modificationTime_; 7341 } 7342 7343 // optional uint64 nsQuota = 2; 7344 public static final int NSQUOTA_FIELD_NUMBER = 2; 7345 private long nsQuota_; 7346 /** 7347 * <code>optional uint64 nsQuota = 2;</code> 7348 * 7349 * <pre> 7350 * namespace quota 7351 * </pre> 7352 */ 7353 public boolean hasNsQuota() { 7354 return ((bitField0_ & 0x00000002) == 0x00000002); 7355 } 7356 /** 7357 * <code>optional uint64 nsQuota = 2;</code> 7358 * 7359 * <pre> 7360 * namespace quota 7361 * </pre> 7362 */ 7363 public long getNsQuota() { 7364 return nsQuota_; 7365 } 7366 7367 // optional uint64 dsQuota = 3; 7368 public static final int DSQUOTA_FIELD_NUMBER = 3; 7369 private long dsQuota_; 7370 /** 7371 * <code>optional uint64 dsQuota = 3;</code> 7372 * 7373 * <pre> 7374 * diskspace quota 7375 * </pre> 7376 */ 7377 public boolean hasDsQuota() { 7378 return ((bitField0_ & 0x00000004) == 0x00000004); 7379 } 7380 /** 7381 * <code>optional uint64 dsQuota = 3;</code> 7382 * 7383 * <pre> 7384 * diskspace quota 7385 * </pre> 7386 */ 7387 public long getDsQuota() { 7388 return dsQuota_; 7389 } 7390 7391 // optional fixed64 permission = 4; 7392 public static final int PERMISSION_FIELD_NUMBER = 4; 7393 private long permission_; 7394 /** 7395 * <code>optional fixed64 permission = 4;</code> 7396 */ 7397 public boolean hasPermission() { 7398 return ((bitField0_ & 0x00000008) == 0x00000008); 7399 } 7400 /** 7401 * <code>optional fixed64 permission = 4;</code> 7402 */ 7403 public long getPermission() { 7404 return permission_; 7405 } 7406 7407 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 7408 public static final int ACL_FIELD_NUMBER = 5; 7409 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_; 7410 /** 7411 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7412 */ 7413 public boolean hasAcl() { 7414 return ((bitField0_ & 0x00000010) == 0x00000010); 7415 } 7416 /** 7417 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7418 */ 7419 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 7420 return acl_; 7421 } 7422 /** 7423 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7424 */ 7425 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 7426 return acl_; 7427 } 7428 7429 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6; 7430 public static final int XATTRS_FIELD_NUMBER = 6; 7431 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto xAttrs_; 7432 /** 7433 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7434 */ 7435 public boolean hasXAttrs() { 7436 return ((bitField0_ & 0x00000020) == 0x00000020); 7437 } 7438 /** 7439 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7440 */ 7441 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs() { 7442 return xAttrs_; 7443 } 7444 /** 7445 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 7446 */ 7447 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder() { 7448 return xAttrs_; 7449 } 7450 7451 private void initFields() { 7452 modificationTime_ = 0L; 7453 nsQuota_ = 0L; 7454 dsQuota_ = 0L; 7455 permission_ = 0L; 7456 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 7457 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 7458 } 7459 private byte memoizedIsInitialized = -1; 7460 public final boolean isInitialized() { 7461 byte isInitialized = memoizedIsInitialized; 7462 if (isInitialized != -1) return isInitialized == 1; 7463 7464 if (hasXAttrs()) { 7465 if (!getXAttrs().isInitialized()) { 7466 memoizedIsInitialized = 0; 7467 return false; 7468 } 7469 } 7470 memoizedIsInitialized = 1; 7471 return true; 7472 } 7473 7474 public void writeTo(com.google.protobuf.CodedOutputStream output) 7475 throws java.io.IOException { 7476 getSerializedSize(); 7477 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7478 output.writeUInt64(1, modificationTime_); 7479 } 7480 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7481 output.writeUInt64(2, nsQuota_); 7482 } 7483 if (((bitField0_ & 0x00000004) == 0x00000004)) { 7484 output.writeUInt64(3, dsQuota_); 7485 } 7486 if (((bitField0_ & 0x00000008) == 0x00000008)) { 7487 output.writeFixed64(4, permission_); 7488 } 7489 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7490 output.writeMessage(5, acl_); 7491 } 7492 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7493 output.writeMessage(6, xAttrs_); 7494 } 7495 getUnknownFields().writeTo(output); 7496 } 7497 7498 private int memoizedSerializedSize = -1; 7499 public int getSerializedSize() { 7500 int size = memoizedSerializedSize; 7501 if (size != -1) return size; 7502 7503 size = 0; 7504 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7505 size += com.google.protobuf.CodedOutputStream 7506 .computeUInt64Size(1, modificationTime_); 7507 } 7508 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7509 size += com.google.protobuf.CodedOutputStream 7510 .computeUInt64Size(2, nsQuota_); 7511 } 7512 if (((bitField0_ & 0x00000004) == 0x00000004)) { 7513 size += com.google.protobuf.CodedOutputStream 7514 .computeUInt64Size(3, dsQuota_); 7515 } 7516 if (((bitField0_ & 0x00000008) == 0x00000008)) { 7517 size += com.google.protobuf.CodedOutputStream 7518 .computeFixed64Size(4, permission_); 7519 } 7520 if (((bitField0_ & 0x00000010) == 0x00000010)) { 7521 size += com.google.protobuf.CodedOutputStream 7522 .computeMessageSize(5, acl_); 7523 } 7524 if (((bitField0_ & 0x00000020) == 0x00000020)) { 7525 size += com.google.protobuf.CodedOutputStream 7526 .computeMessageSize(6, xAttrs_); 7527 } 7528 size += getUnknownFields().getSerializedSize(); 7529 memoizedSerializedSize = size; 7530 return size; 7531 } 7532 7533 private static final long serialVersionUID = 0L; 7534 @java.lang.Override 7535 protected java.lang.Object writeReplace() 7536 throws java.io.ObjectStreamException { 7537 return super.writeReplace(); 7538 } 7539 7540 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7541 com.google.protobuf.ByteString data) 7542 throws com.google.protobuf.InvalidProtocolBufferException { 7543 return PARSER.parseFrom(data); 7544 } 7545 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7546 com.google.protobuf.ByteString data, 7547 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7548 throws com.google.protobuf.InvalidProtocolBufferException { 7549 return PARSER.parseFrom(data, extensionRegistry); 7550 } 7551 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom(byte[] data) 7552 throws com.google.protobuf.InvalidProtocolBufferException { 7553 return PARSER.parseFrom(data); 7554 } 7555 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7556 byte[] data, 7557 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7558 throws com.google.protobuf.InvalidProtocolBufferException { 7559 return PARSER.parseFrom(data, extensionRegistry); 7560 } 7561 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom(java.io.InputStream input) 7562 throws java.io.IOException { 7563 return PARSER.parseFrom(input); 7564 } 7565 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7566 java.io.InputStream input, 7567 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7568 throws java.io.IOException { 7569 return PARSER.parseFrom(input, extensionRegistry); 7570 } 7571 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseDelimitedFrom(java.io.InputStream input) 7572 throws java.io.IOException { 7573 return PARSER.parseDelimitedFrom(input); 7574 } 7575 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseDelimitedFrom( 7576 java.io.InputStream input, 7577 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7578 throws java.io.IOException { 7579 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7580 } 7581 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7582 com.google.protobuf.CodedInputStream input) 7583 throws java.io.IOException { 7584 return PARSER.parseFrom(input); 7585 } 7586 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parseFrom( 7587 com.google.protobuf.CodedInputStream input, 7588 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7589 throws java.io.IOException { 7590 return PARSER.parseFrom(input, extensionRegistry); 7591 } 7592 7593 public static Builder newBuilder() { return Builder.create(); } 7594 public Builder newBuilderForType() { return newBuilder(); } 7595 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory prototype) { 7596 return newBuilder().mergeFrom(prototype); 7597 } 7598 public Builder toBuilder() { return newBuilder(this); } 7599 7600 @java.lang.Override 7601 protected Builder newBuilderForType( 7602 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7603 Builder builder = new Builder(parent); 7604 return builder; 7605 } 7606 /** 7607 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeDirectory} 7608 */ 7609 public static final class Builder extends 7610 com.google.protobuf.GeneratedMessage.Builder<Builder> 7611 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder { 7612 public static final com.google.protobuf.Descriptors.Descriptor 7613 getDescriptor() { 7614 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 7615 } 7616 7617 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 7618 internalGetFieldAccessorTable() { 7619 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable 7620 .ensureFieldAccessorsInitialized( 7621 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder.class); 7622 } 7623 7624 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder() 7625 private Builder() { 7626 maybeForceBuilderInitialization(); 7627 } 7628 7629 private Builder( 7630 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7631 super(parent); 7632 maybeForceBuilderInitialization(); 7633 } 7634 private void maybeForceBuilderInitialization() { 7635 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7636 getAclFieldBuilder(); 7637 getXAttrsFieldBuilder(); 7638 } 7639 } 7640 private static Builder create() { 7641 return new Builder(); 7642 } 7643 7644 public Builder clear() { 7645 super.clear(); 7646 modificationTime_ = 0L; 7647 bitField0_ = (bitField0_ & ~0x00000001); 7648 nsQuota_ = 0L; 7649 bitField0_ = (bitField0_ & ~0x00000002); 7650 dsQuota_ = 0L; 7651 bitField0_ = (bitField0_ & ~0x00000004); 7652 permission_ = 0L; 7653 bitField0_ = (bitField0_ & ~0x00000008); 7654 if (aclBuilder_ == null) { 7655 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 7656 } else { 7657 aclBuilder_.clear(); 7658 } 7659 bitField0_ = (bitField0_ & ~0x00000010); 7660 if (xAttrsBuilder_ == null) { 7661 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 7662 } else { 7663 xAttrsBuilder_.clear(); 7664 } 7665 bitField0_ = (bitField0_ & ~0x00000020); 7666 return this; 7667 } 7668 7669 public Builder clone() { 7670 return create().mergeFrom(buildPartial()); 7671 } 7672 7673 public com.google.protobuf.Descriptors.Descriptor 7674 getDescriptorForType() { 7675 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 7676 } 7677 7678 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDefaultInstanceForType() { 7679 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 7680 } 7681 7682 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory build() { 7683 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory result = buildPartial(); 7684 if (!result.isInitialized()) { 7685 throw newUninitializedMessageException(result); 7686 } 7687 return result; 7688 } 7689 7690 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory buildPartial() { 7691 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory(this); 7692 int from_bitField0_ = bitField0_; 7693 int to_bitField0_ = 0; 7694 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 7695 to_bitField0_ |= 0x00000001; 7696 } 7697 result.modificationTime_ = modificationTime_; 7698 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 7699 to_bitField0_ |= 0x00000002; 7700 } 7701 result.nsQuota_ = nsQuota_; 7702 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 7703 to_bitField0_ |= 0x00000004; 7704 } 7705 result.dsQuota_ = dsQuota_; 7706 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 7707 to_bitField0_ |= 0x00000008; 7708 } 7709 result.permission_ = permission_; 7710 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 7711 to_bitField0_ |= 0x00000010; 7712 } 7713 if (aclBuilder_ == null) { 7714 result.acl_ = acl_; 7715 } else { 7716 result.acl_ = aclBuilder_.build(); 7717 } 7718 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 7719 to_bitField0_ |= 0x00000020; 7720 } 7721 if (xAttrsBuilder_ == null) { 7722 result.xAttrs_ = xAttrs_; 7723 } else { 7724 result.xAttrs_ = xAttrsBuilder_.build(); 7725 } 7726 result.bitField0_ = to_bitField0_; 7727 onBuilt(); 7728 return result; 7729 } 7730 7731 public Builder mergeFrom(com.google.protobuf.Message other) { 7732 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory) { 7733 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory)other); 7734 } else { 7735 super.mergeFrom(other); 7736 return this; 7737 } 7738 } 7739 7740 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory other) { 7741 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) return this; 7742 if (other.hasModificationTime()) { 7743 setModificationTime(other.getModificationTime()); 7744 } 7745 if (other.hasNsQuota()) { 7746 setNsQuota(other.getNsQuota()); 7747 } 7748 if (other.hasDsQuota()) { 7749 setDsQuota(other.getDsQuota()); 7750 } 7751 if (other.hasPermission()) { 7752 setPermission(other.getPermission()); 7753 } 7754 if (other.hasAcl()) { 7755 mergeAcl(other.getAcl()); 7756 } 7757 if (other.hasXAttrs()) { 7758 mergeXAttrs(other.getXAttrs()); 7759 } 7760 this.mergeUnknownFields(other.getUnknownFields()); 7761 return this; 7762 } 7763 7764 public final boolean isInitialized() { 7765 if (hasXAttrs()) { 7766 if (!getXAttrs().isInitialized()) { 7767 7768 return false; 7769 } 7770 } 7771 return true; 7772 } 7773 7774 public Builder mergeFrom( 7775 com.google.protobuf.CodedInputStream input, 7776 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7777 throws java.io.IOException { 7778 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory parsedMessage = null; 7779 try { 7780 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7781 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7782 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory) e.getUnfinishedMessage(); 7783 throw e; 7784 } finally { 7785 if (parsedMessage != null) { 7786 mergeFrom(parsedMessage); 7787 } 7788 } 7789 return this; 7790 } 7791 private int bitField0_; 7792 7793 // optional uint64 modificationTime = 1; 7794 private long modificationTime_ ; 7795 /** 7796 * <code>optional uint64 modificationTime = 1;</code> 7797 */ 7798 public boolean hasModificationTime() { 7799 return ((bitField0_ & 0x00000001) == 0x00000001); 7800 } 7801 /** 7802 * <code>optional uint64 modificationTime = 1;</code> 7803 */ 7804 public long getModificationTime() { 7805 return modificationTime_; 7806 } 7807 /** 7808 * <code>optional uint64 modificationTime = 1;</code> 7809 */ 7810 public Builder setModificationTime(long value) { 7811 bitField0_ |= 0x00000001; 7812 modificationTime_ = value; 7813 onChanged(); 7814 return this; 7815 } 7816 /** 7817 * <code>optional uint64 modificationTime = 1;</code> 7818 */ 7819 public Builder clearModificationTime() { 7820 bitField0_ = (bitField0_ & ~0x00000001); 7821 modificationTime_ = 0L; 7822 onChanged(); 7823 return this; 7824 } 7825 7826 // optional uint64 nsQuota = 2; 7827 private long nsQuota_ ; 7828 /** 7829 * <code>optional uint64 nsQuota = 2;</code> 7830 * 7831 * <pre> 7832 * namespace quota 7833 * </pre> 7834 */ 7835 public boolean hasNsQuota() { 7836 return ((bitField0_ & 0x00000002) == 0x00000002); 7837 } 7838 /** 7839 * <code>optional uint64 nsQuota = 2;</code> 7840 * 7841 * <pre> 7842 * namespace quota 7843 * </pre> 7844 */ 7845 public long getNsQuota() { 7846 return nsQuota_; 7847 } 7848 /** 7849 * <code>optional uint64 nsQuota = 2;</code> 7850 * 7851 * <pre> 7852 * namespace quota 7853 * </pre> 7854 */ 7855 public Builder setNsQuota(long value) { 7856 bitField0_ |= 0x00000002; 7857 nsQuota_ = value; 7858 onChanged(); 7859 return this; 7860 } 7861 /** 7862 * <code>optional uint64 nsQuota = 2;</code> 7863 * 7864 * <pre> 7865 * namespace quota 7866 * </pre> 7867 */ 7868 public Builder clearNsQuota() { 7869 bitField0_ = (bitField0_ & ~0x00000002); 7870 nsQuota_ = 0L; 7871 onChanged(); 7872 return this; 7873 } 7874 7875 // optional uint64 dsQuota = 3; 7876 private long dsQuota_ ; 7877 /** 7878 * <code>optional uint64 dsQuota = 3;</code> 7879 * 7880 * <pre> 7881 * diskspace quota 7882 * </pre> 7883 */ 7884 public boolean hasDsQuota() { 7885 return ((bitField0_ & 0x00000004) == 0x00000004); 7886 } 7887 /** 7888 * <code>optional uint64 dsQuota = 3;</code> 7889 * 7890 * <pre> 7891 * diskspace quota 7892 * </pre> 7893 */ 7894 public long getDsQuota() { 7895 return dsQuota_; 7896 } 7897 /** 7898 * <code>optional uint64 dsQuota = 3;</code> 7899 * 7900 * <pre> 7901 * diskspace quota 7902 * </pre> 7903 */ 7904 public Builder setDsQuota(long value) { 7905 bitField0_ |= 0x00000004; 7906 dsQuota_ = value; 7907 onChanged(); 7908 return this; 7909 } 7910 /** 7911 * <code>optional uint64 dsQuota = 3;</code> 7912 * 7913 * <pre> 7914 * diskspace quota 7915 * </pre> 7916 */ 7917 public Builder clearDsQuota() { 7918 bitField0_ = (bitField0_ & ~0x00000004); 7919 dsQuota_ = 0L; 7920 onChanged(); 7921 return this; 7922 } 7923 7924 // optional fixed64 permission = 4; 7925 private long permission_ ; 7926 /** 7927 * <code>optional fixed64 permission = 4;</code> 7928 */ 7929 public boolean hasPermission() { 7930 return ((bitField0_ & 0x00000008) == 0x00000008); 7931 } 7932 /** 7933 * <code>optional fixed64 permission = 4;</code> 7934 */ 7935 public long getPermission() { 7936 return permission_; 7937 } 7938 /** 7939 * <code>optional fixed64 permission = 4;</code> 7940 */ 7941 public Builder setPermission(long value) { 7942 bitField0_ |= 0x00000008; 7943 permission_ = value; 7944 onChanged(); 7945 return this; 7946 } 7947 /** 7948 * <code>optional fixed64 permission = 4;</code> 7949 */ 7950 public Builder clearPermission() { 7951 bitField0_ = (bitField0_ & ~0x00000008); 7952 permission_ = 0L; 7953 onChanged(); 7954 return this; 7955 } 7956 7957 // optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5; 7958 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 7959 private com.google.protobuf.SingleFieldBuilder< 7960 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> aclBuilder_; 7961 /** 7962 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7963 */ 7964 public boolean hasAcl() { 7965 return ((bitField0_ & 0x00000010) == 0x00000010); 7966 } 7967 /** 7968 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7969 */ 7970 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto getAcl() { 7971 if (aclBuilder_ == null) { 7972 return acl_; 7973 } else { 7974 return aclBuilder_.getMessage(); 7975 } 7976 } 7977 /** 7978 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7979 */ 7980 public Builder setAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 7981 if (aclBuilder_ == null) { 7982 if (value == null) { 7983 throw new NullPointerException(); 7984 } 7985 acl_ = value; 7986 onChanged(); 7987 } else { 7988 aclBuilder_.setMessage(value); 7989 } 7990 bitField0_ |= 0x00000010; 7991 return this; 7992 } 7993 /** 7994 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 7995 */ 7996 public Builder setAcl( 7997 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder builderForValue) { 7998 if (aclBuilder_ == null) { 7999 acl_ = builderForValue.build(); 8000 onChanged(); 8001 } else { 8002 aclBuilder_.setMessage(builderForValue.build()); 8003 } 8004 bitField0_ |= 0x00000010; 8005 return this; 8006 } 8007 /** 8008 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 8009 */ 8010 public Builder mergeAcl(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto value) { 8011 if (aclBuilder_ == null) { 8012 if (((bitField0_ & 0x00000010) == 0x00000010) && 8013 acl_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance()) { 8014 acl_ = 8015 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.newBuilder(acl_).mergeFrom(value).buildPartial(); 8016 } else { 8017 acl_ = value; 8018 } 8019 onChanged(); 8020 } else { 8021 aclBuilder_.mergeFrom(value); 8022 } 8023 bitField0_ |= 0x00000010; 8024 return this; 8025 } 8026 /** 8027 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 8028 */ 8029 public Builder clearAcl() { 8030 if (aclBuilder_ == null) { 8031 acl_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.getDefaultInstance(); 8032 onChanged(); 8033 } else { 8034 aclBuilder_.clear(); 8035 } 8036 bitField0_ = (bitField0_ & ~0x00000010); 8037 return this; 8038 } 8039 /** 8040 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 8041 */ 8042 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder getAclBuilder() { 8043 bitField0_ |= 0x00000010; 8044 onChanged(); 8045 return getAclFieldBuilder().getBuilder(); 8046 } 8047 /** 8048 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 8049 */ 8050 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder getAclOrBuilder() { 8051 if (aclBuilder_ != null) { 8052 return aclBuilder_.getMessageOrBuilder(); 8053 } else { 8054 return acl_; 8055 } 8056 } 8057 /** 8058 * <code>optional .hadoop.hdfs.fsimage.INodeSection.AclFeatureProto acl = 5;</code> 8059 */ 8060 private com.google.protobuf.SingleFieldBuilder< 8061 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder> 8062 getAclFieldBuilder() { 8063 if (aclBuilder_ == null) { 8064 aclBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8065 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.AclFeatureProtoOrBuilder>( 8066 acl_, 8067 getParentForChildren(), 8068 isClean()); 8069 acl_ = null; 8070 } 8071 return aclBuilder_; 8072 } 8073 8074 // optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6; 8075 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 8076 private com.google.protobuf.SingleFieldBuilder< 8077 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder> xAttrsBuilder_; 8078 /** 8079 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8080 */ 8081 public boolean hasXAttrs() { 8082 return ((bitField0_ & 0x00000020) == 0x00000020); 8083 } 8084 /** 8085 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8086 */ 8087 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto getXAttrs() { 8088 if (xAttrsBuilder_ == null) { 8089 return xAttrs_; 8090 } else { 8091 return xAttrsBuilder_.getMessage(); 8092 } 8093 } 8094 /** 8095 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8096 */ 8097 public Builder setXAttrs(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto value) { 8098 if (xAttrsBuilder_ == null) { 8099 if (value == null) { 8100 throw new NullPointerException(); 8101 } 8102 xAttrs_ = value; 8103 onChanged(); 8104 } else { 8105 xAttrsBuilder_.setMessage(value); 8106 } 8107 bitField0_ |= 0x00000020; 8108 return this; 8109 } 8110 /** 8111 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8112 */ 8113 public Builder setXAttrs( 8114 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder builderForValue) { 8115 if (xAttrsBuilder_ == null) { 8116 xAttrs_ = builderForValue.build(); 8117 onChanged(); 8118 } else { 8119 xAttrsBuilder_.setMessage(builderForValue.build()); 8120 } 8121 bitField0_ |= 0x00000020; 8122 return this; 8123 } 8124 /** 8125 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8126 */ 8127 public Builder mergeXAttrs(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto value) { 8128 if (xAttrsBuilder_ == null) { 8129 if (((bitField0_ & 0x00000020) == 0x00000020) && 8130 xAttrs_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance()) { 8131 xAttrs_ = 8132 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.newBuilder(xAttrs_).mergeFrom(value).buildPartial(); 8133 } else { 8134 xAttrs_ = value; 8135 } 8136 onChanged(); 8137 } else { 8138 xAttrsBuilder_.mergeFrom(value); 8139 } 8140 bitField0_ |= 0x00000020; 8141 return this; 8142 } 8143 /** 8144 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8145 */ 8146 public Builder clearXAttrs() { 8147 if (xAttrsBuilder_ == null) { 8148 xAttrs_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.getDefaultInstance(); 8149 onChanged(); 8150 } else { 8151 xAttrsBuilder_.clear(); 8152 } 8153 bitField0_ = (bitField0_ & ~0x00000020); 8154 return this; 8155 } 8156 /** 8157 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8158 */ 8159 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder getXAttrsBuilder() { 8160 bitField0_ |= 0x00000020; 8161 onChanged(); 8162 return getXAttrsFieldBuilder().getBuilder(); 8163 } 8164 /** 8165 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8166 */ 8167 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder getXAttrsOrBuilder() { 8168 if (xAttrsBuilder_ != null) { 8169 return xAttrsBuilder_.getMessageOrBuilder(); 8170 } else { 8171 return xAttrs_; 8172 } 8173 } 8174 /** 8175 * <code>optional .hadoop.hdfs.fsimage.INodeSection.XAttrFeatureProto xAttrs = 6;</code> 8176 */ 8177 private com.google.protobuf.SingleFieldBuilder< 8178 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder> 8179 getXAttrsFieldBuilder() { 8180 if (xAttrsBuilder_ == null) { 8181 xAttrsBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8182 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProto.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.XAttrFeatureProtoOrBuilder>( 8183 xAttrs_, 8184 getParentForChildren(), 8185 isClean()); 8186 xAttrs_ = null; 8187 } 8188 return xAttrsBuilder_; 8189 } 8190 8191 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeDirectory) 8192 } 8193 8194 static { 8195 defaultInstance = new INodeDirectory(true); 8196 defaultInstance.initFields(); 8197 } 8198 8199 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeDirectory) 8200 } 8201 8202 public interface INodeSymlinkOrBuilder 8203 extends com.google.protobuf.MessageOrBuilder { 8204 8205 // optional fixed64 permission = 1; 8206 /** 8207 * <code>optional fixed64 permission = 1;</code> 8208 */ 8209 boolean hasPermission(); 8210 /** 8211 * <code>optional fixed64 permission = 1;</code> 8212 */ 8213 long getPermission(); 8214 8215 // optional bytes target = 2; 8216 /** 8217 * <code>optional bytes target = 2;</code> 8218 */ 8219 boolean hasTarget(); 8220 /** 8221 * <code>optional bytes target = 2;</code> 8222 */ 8223 com.google.protobuf.ByteString getTarget(); 8224 8225 // optional uint64 modificationTime = 3; 8226 /** 8227 * <code>optional uint64 modificationTime = 3;</code> 8228 */ 8229 boolean hasModificationTime(); 8230 /** 8231 * <code>optional uint64 modificationTime = 3;</code> 8232 */ 8233 long getModificationTime(); 8234 8235 // optional uint64 accessTime = 4; 8236 /** 8237 * <code>optional uint64 accessTime = 4;</code> 8238 */ 8239 boolean hasAccessTime(); 8240 /** 8241 * <code>optional uint64 accessTime = 4;</code> 8242 */ 8243 long getAccessTime(); 8244 } 8245 /** 8246 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeSymlink} 8247 */ 8248 public static final class INodeSymlink extends 8249 com.google.protobuf.GeneratedMessage 8250 implements INodeSymlinkOrBuilder { 8251 // Use INodeSymlink.newBuilder() to construct. 8252 private INodeSymlink(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8253 super(builder); 8254 this.unknownFields = builder.getUnknownFields(); 8255 } 8256 private INodeSymlink(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8257 8258 private static final INodeSymlink defaultInstance; 8259 public static INodeSymlink getDefaultInstance() { 8260 return defaultInstance; 8261 } 8262 8263 public INodeSymlink getDefaultInstanceForType() { 8264 return defaultInstance; 8265 } 8266 8267 private final com.google.protobuf.UnknownFieldSet unknownFields; 8268 @java.lang.Override 8269 public final com.google.protobuf.UnknownFieldSet 8270 getUnknownFields() { 8271 return this.unknownFields; 8272 } 8273 private INodeSymlink( 8274 com.google.protobuf.CodedInputStream input, 8275 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8276 throws com.google.protobuf.InvalidProtocolBufferException { 8277 initFields(); 8278 int mutable_bitField0_ = 0; 8279 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8280 com.google.protobuf.UnknownFieldSet.newBuilder(); 8281 try { 8282 boolean done = false; 8283 while (!done) { 8284 int tag = input.readTag(); 8285 switch (tag) { 8286 case 0: 8287 done = true; 8288 break; 8289 default: { 8290 if (!parseUnknownField(input, unknownFields, 8291 extensionRegistry, tag)) { 8292 done = true; 8293 } 8294 break; 8295 } 8296 case 9: { 8297 bitField0_ |= 0x00000001; 8298 permission_ = input.readFixed64(); 8299 break; 8300 } 8301 case 18: { 8302 bitField0_ |= 0x00000002; 8303 target_ = input.readBytes(); 8304 break; 8305 } 8306 case 24: { 8307 bitField0_ |= 0x00000004; 8308 modificationTime_ = input.readUInt64(); 8309 break; 8310 } 8311 case 32: { 8312 bitField0_ |= 0x00000008; 8313 accessTime_ = input.readUInt64(); 8314 break; 8315 } 8316 } 8317 } 8318 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8319 throw e.setUnfinishedMessage(this); 8320 } catch (java.io.IOException e) { 8321 throw new com.google.protobuf.InvalidProtocolBufferException( 8322 e.getMessage()).setUnfinishedMessage(this); 8323 } finally { 8324 this.unknownFields = unknownFields.build(); 8325 makeExtensionsImmutable(); 8326 } 8327 } 8328 public static final com.google.protobuf.Descriptors.Descriptor 8329 getDescriptor() { 8330 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 8331 } 8332 8333 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8334 internalGetFieldAccessorTable() { 8335 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable 8336 .ensureFieldAccessorsInitialized( 8337 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder.class); 8338 } 8339 8340 public static com.google.protobuf.Parser<INodeSymlink> PARSER = 8341 new com.google.protobuf.AbstractParser<INodeSymlink>() { 8342 public INodeSymlink parsePartialFrom( 8343 com.google.protobuf.CodedInputStream input, 8344 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8345 throws com.google.protobuf.InvalidProtocolBufferException { 8346 return new INodeSymlink(input, extensionRegistry); 8347 } 8348 }; 8349 8350 @java.lang.Override 8351 public com.google.protobuf.Parser<INodeSymlink> getParserForType() { 8352 return PARSER; 8353 } 8354 8355 private int bitField0_; 8356 // optional fixed64 permission = 1; 8357 public static final int PERMISSION_FIELD_NUMBER = 1; 8358 private long permission_; 8359 /** 8360 * <code>optional fixed64 permission = 1;</code> 8361 */ 8362 public boolean hasPermission() { 8363 return ((bitField0_ & 0x00000001) == 0x00000001); 8364 } 8365 /** 8366 * <code>optional fixed64 permission = 1;</code> 8367 */ 8368 public long getPermission() { 8369 return permission_; 8370 } 8371 8372 // optional bytes target = 2; 8373 public static final int TARGET_FIELD_NUMBER = 2; 8374 private com.google.protobuf.ByteString target_; 8375 /** 8376 * <code>optional bytes target = 2;</code> 8377 */ 8378 public boolean hasTarget() { 8379 return ((bitField0_ & 0x00000002) == 0x00000002); 8380 } 8381 /** 8382 * <code>optional bytes target = 2;</code> 8383 */ 8384 public com.google.protobuf.ByteString getTarget() { 8385 return target_; 8386 } 8387 8388 // optional uint64 modificationTime = 3; 8389 public static final int MODIFICATIONTIME_FIELD_NUMBER = 3; 8390 private long modificationTime_; 8391 /** 8392 * <code>optional uint64 modificationTime = 3;</code> 8393 */ 8394 public boolean hasModificationTime() { 8395 return ((bitField0_ & 0x00000004) == 0x00000004); 8396 } 8397 /** 8398 * <code>optional uint64 modificationTime = 3;</code> 8399 */ 8400 public long getModificationTime() { 8401 return modificationTime_; 8402 } 8403 8404 // optional uint64 accessTime = 4; 8405 public static final int ACCESSTIME_FIELD_NUMBER = 4; 8406 private long accessTime_; 8407 /** 8408 * <code>optional uint64 accessTime = 4;</code> 8409 */ 8410 public boolean hasAccessTime() { 8411 return ((bitField0_ & 0x00000008) == 0x00000008); 8412 } 8413 /** 8414 * <code>optional uint64 accessTime = 4;</code> 8415 */ 8416 public long getAccessTime() { 8417 return accessTime_; 8418 } 8419 8420 private void initFields() { 8421 permission_ = 0L; 8422 target_ = com.google.protobuf.ByteString.EMPTY; 8423 modificationTime_ = 0L; 8424 accessTime_ = 0L; 8425 } 8426 private byte memoizedIsInitialized = -1; 8427 public final boolean isInitialized() { 8428 byte isInitialized = memoizedIsInitialized; 8429 if (isInitialized != -1) return isInitialized == 1; 8430 8431 memoizedIsInitialized = 1; 8432 return true; 8433 } 8434 8435 public void writeTo(com.google.protobuf.CodedOutputStream output) 8436 throws java.io.IOException { 8437 getSerializedSize(); 8438 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8439 output.writeFixed64(1, permission_); 8440 } 8441 if (((bitField0_ & 0x00000002) == 0x00000002)) { 8442 output.writeBytes(2, target_); 8443 } 8444 if (((bitField0_ & 0x00000004) == 0x00000004)) { 8445 output.writeUInt64(3, modificationTime_); 8446 } 8447 if (((bitField0_ & 0x00000008) == 0x00000008)) { 8448 output.writeUInt64(4, accessTime_); 8449 } 8450 getUnknownFields().writeTo(output); 8451 } 8452 8453 private int memoizedSerializedSize = -1; 8454 public int getSerializedSize() { 8455 int size = memoizedSerializedSize; 8456 if (size != -1) return size; 8457 8458 size = 0; 8459 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8460 size += com.google.protobuf.CodedOutputStream 8461 .computeFixed64Size(1, permission_); 8462 } 8463 if (((bitField0_ & 0x00000002) == 0x00000002)) { 8464 size += com.google.protobuf.CodedOutputStream 8465 .computeBytesSize(2, target_); 8466 } 8467 if (((bitField0_ & 0x00000004) == 0x00000004)) { 8468 size += com.google.protobuf.CodedOutputStream 8469 .computeUInt64Size(3, modificationTime_); 8470 } 8471 if (((bitField0_ & 0x00000008) == 0x00000008)) { 8472 size += com.google.protobuf.CodedOutputStream 8473 .computeUInt64Size(4, accessTime_); 8474 } 8475 size += getUnknownFields().getSerializedSize(); 8476 memoizedSerializedSize = size; 8477 return size; 8478 } 8479 8480 private static final long serialVersionUID = 0L; 8481 @java.lang.Override 8482 protected java.lang.Object writeReplace() 8483 throws java.io.ObjectStreamException { 8484 return super.writeReplace(); 8485 } 8486 8487 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8488 com.google.protobuf.ByteString data) 8489 throws com.google.protobuf.InvalidProtocolBufferException { 8490 return PARSER.parseFrom(data); 8491 } 8492 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8493 com.google.protobuf.ByteString data, 8494 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8495 throws com.google.protobuf.InvalidProtocolBufferException { 8496 return PARSER.parseFrom(data, extensionRegistry); 8497 } 8498 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom(byte[] data) 8499 throws com.google.protobuf.InvalidProtocolBufferException { 8500 return PARSER.parseFrom(data); 8501 } 8502 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8503 byte[] data, 8504 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8505 throws com.google.protobuf.InvalidProtocolBufferException { 8506 return PARSER.parseFrom(data, extensionRegistry); 8507 } 8508 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom(java.io.InputStream input) 8509 throws java.io.IOException { 8510 return PARSER.parseFrom(input); 8511 } 8512 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8513 java.io.InputStream input, 8514 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8515 throws java.io.IOException { 8516 return PARSER.parseFrom(input, extensionRegistry); 8517 } 8518 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseDelimitedFrom(java.io.InputStream input) 8519 throws java.io.IOException { 8520 return PARSER.parseDelimitedFrom(input); 8521 } 8522 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseDelimitedFrom( 8523 java.io.InputStream input, 8524 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8525 throws java.io.IOException { 8526 return PARSER.parseDelimitedFrom(input, extensionRegistry); 8527 } 8528 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8529 com.google.protobuf.CodedInputStream input) 8530 throws java.io.IOException { 8531 return PARSER.parseFrom(input); 8532 } 8533 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parseFrom( 8534 com.google.protobuf.CodedInputStream input, 8535 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8536 throws java.io.IOException { 8537 return PARSER.parseFrom(input, extensionRegistry); 8538 } 8539 8540 public static Builder newBuilder() { return Builder.create(); } 8541 public Builder newBuilderForType() { return newBuilder(); } 8542 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink prototype) { 8543 return newBuilder().mergeFrom(prototype); 8544 } 8545 public Builder toBuilder() { return newBuilder(this); } 8546 8547 @java.lang.Override 8548 protected Builder newBuilderForType( 8549 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8550 Builder builder = new Builder(parent); 8551 return builder; 8552 } 8553 /** 8554 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INodeSymlink} 8555 */ 8556 public static final class Builder extends 8557 com.google.protobuf.GeneratedMessage.Builder<Builder> 8558 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder { 8559 public static final com.google.protobuf.Descriptors.Descriptor 8560 getDescriptor() { 8561 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 8562 } 8563 8564 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 8565 internalGetFieldAccessorTable() { 8566 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable 8567 .ensureFieldAccessorsInitialized( 8568 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder.class); 8569 } 8570 8571 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.newBuilder() 8572 private Builder() { 8573 maybeForceBuilderInitialization(); 8574 } 8575 8576 private Builder( 8577 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8578 super(parent); 8579 maybeForceBuilderInitialization(); 8580 } 8581 private void maybeForceBuilderInitialization() { 8582 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8583 } 8584 } 8585 private static Builder create() { 8586 return new Builder(); 8587 } 8588 8589 public Builder clear() { 8590 super.clear(); 8591 permission_ = 0L; 8592 bitField0_ = (bitField0_ & ~0x00000001); 8593 target_ = com.google.protobuf.ByteString.EMPTY; 8594 bitField0_ = (bitField0_ & ~0x00000002); 8595 modificationTime_ = 0L; 8596 bitField0_ = (bitField0_ & ~0x00000004); 8597 accessTime_ = 0L; 8598 bitField0_ = (bitField0_ & ~0x00000008); 8599 return this; 8600 } 8601 8602 public Builder clone() { 8603 return create().mergeFrom(buildPartial()); 8604 } 8605 8606 public com.google.protobuf.Descriptors.Descriptor 8607 getDescriptorForType() { 8608 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 8609 } 8610 8611 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getDefaultInstanceForType() { 8612 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 8613 } 8614 8615 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink build() { 8616 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink result = buildPartial(); 8617 if (!result.isInitialized()) { 8618 throw newUninitializedMessageException(result); 8619 } 8620 return result; 8621 } 8622 8623 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink buildPartial() { 8624 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink(this); 8625 int from_bitField0_ = bitField0_; 8626 int to_bitField0_ = 0; 8627 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 8628 to_bitField0_ |= 0x00000001; 8629 } 8630 result.permission_ = permission_; 8631 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 8632 to_bitField0_ |= 0x00000002; 8633 } 8634 result.target_ = target_; 8635 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 8636 to_bitField0_ |= 0x00000004; 8637 } 8638 result.modificationTime_ = modificationTime_; 8639 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 8640 to_bitField0_ |= 0x00000008; 8641 } 8642 result.accessTime_ = accessTime_; 8643 result.bitField0_ = to_bitField0_; 8644 onBuilt(); 8645 return result; 8646 } 8647 8648 public Builder mergeFrom(com.google.protobuf.Message other) { 8649 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink) { 8650 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink)other); 8651 } else { 8652 super.mergeFrom(other); 8653 return this; 8654 } 8655 } 8656 8657 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink other) { 8658 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance()) return this; 8659 if (other.hasPermission()) { 8660 setPermission(other.getPermission()); 8661 } 8662 if (other.hasTarget()) { 8663 setTarget(other.getTarget()); 8664 } 8665 if (other.hasModificationTime()) { 8666 setModificationTime(other.getModificationTime()); 8667 } 8668 if (other.hasAccessTime()) { 8669 setAccessTime(other.getAccessTime()); 8670 } 8671 this.mergeUnknownFields(other.getUnknownFields()); 8672 return this; 8673 } 8674 8675 public final boolean isInitialized() { 8676 return true; 8677 } 8678 8679 public Builder mergeFrom( 8680 com.google.protobuf.CodedInputStream input, 8681 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8682 throws java.io.IOException { 8683 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink parsedMessage = null; 8684 try { 8685 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 8686 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8687 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink) e.getUnfinishedMessage(); 8688 throw e; 8689 } finally { 8690 if (parsedMessage != null) { 8691 mergeFrom(parsedMessage); 8692 } 8693 } 8694 return this; 8695 } 8696 private int bitField0_; 8697 8698 // optional fixed64 permission = 1; 8699 private long permission_ ; 8700 /** 8701 * <code>optional fixed64 permission = 1;</code> 8702 */ 8703 public boolean hasPermission() { 8704 return ((bitField0_ & 0x00000001) == 0x00000001); 8705 } 8706 /** 8707 * <code>optional fixed64 permission = 1;</code> 8708 */ 8709 public long getPermission() { 8710 return permission_; 8711 } 8712 /** 8713 * <code>optional fixed64 permission = 1;</code> 8714 */ 8715 public Builder setPermission(long value) { 8716 bitField0_ |= 0x00000001; 8717 permission_ = value; 8718 onChanged(); 8719 return this; 8720 } 8721 /** 8722 * <code>optional fixed64 permission = 1;</code> 8723 */ 8724 public Builder clearPermission() { 8725 bitField0_ = (bitField0_ & ~0x00000001); 8726 permission_ = 0L; 8727 onChanged(); 8728 return this; 8729 } 8730 8731 // optional bytes target = 2; 8732 private com.google.protobuf.ByteString target_ = com.google.protobuf.ByteString.EMPTY; 8733 /** 8734 * <code>optional bytes target = 2;</code> 8735 */ 8736 public boolean hasTarget() { 8737 return ((bitField0_ & 0x00000002) == 0x00000002); 8738 } 8739 /** 8740 * <code>optional bytes target = 2;</code> 8741 */ 8742 public com.google.protobuf.ByteString getTarget() { 8743 return target_; 8744 } 8745 /** 8746 * <code>optional bytes target = 2;</code> 8747 */ 8748 public Builder setTarget(com.google.protobuf.ByteString value) { 8749 if (value == null) { 8750 throw new NullPointerException(); 8751 } 8752 bitField0_ |= 0x00000002; 8753 target_ = value; 8754 onChanged(); 8755 return this; 8756 } 8757 /** 8758 * <code>optional bytes target = 2;</code> 8759 */ 8760 public Builder clearTarget() { 8761 bitField0_ = (bitField0_ & ~0x00000002); 8762 target_ = getDefaultInstance().getTarget(); 8763 onChanged(); 8764 return this; 8765 } 8766 8767 // optional uint64 modificationTime = 3; 8768 private long modificationTime_ ; 8769 /** 8770 * <code>optional uint64 modificationTime = 3;</code> 8771 */ 8772 public boolean hasModificationTime() { 8773 return ((bitField0_ & 0x00000004) == 0x00000004); 8774 } 8775 /** 8776 * <code>optional uint64 modificationTime = 3;</code> 8777 */ 8778 public long getModificationTime() { 8779 return modificationTime_; 8780 } 8781 /** 8782 * <code>optional uint64 modificationTime = 3;</code> 8783 */ 8784 public Builder setModificationTime(long value) { 8785 bitField0_ |= 0x00000004; 8786 modificationTime_ = value; 8787 onChanged(); 8788 return this; 8789 } 8790 /** 8791 * <code>optional uint64 modificationTime = 3;</code> 8792 */ 8793 public Builder clearModificationTime() { 8794 bitField0_ = (bitField0_ & ~0x00000004); 8795 modificationTime_ = 0L; 8796 onChanged(); 8797 return this; 8798 } 8799 8800 // optional uint64 accessTime = 4; 8801 private long accessTime_ ; 8802 /** 8803 * <code>optional uint64 accessTime = 4;</code> 8804 */ 8805 public boolean hasAccessTime() { 8806 return ((bitField0_ & 0x00000008) == 0x00000008); 8807 } 8808 /** 8809 * <code>optional uint64 accessTime = 4;</code> 8810 */ 8811 public long getAccessTime() { 8812 return accessTime_; 8813 } 8814 /** 8815 * <code>optional uint64 accessTime = 4;</code> 8816 */ 8817 public Builder setAccessTime(long value) { 8818 bitField0_ |= 0x00000008; 8819 accessTime_ = value; 8820 onChanged(); 8821 return this; 8822 } 8823 /** 8824 * <code>optional uint64 accessTime = 4;</code> 8825 */ 8826 public Builder clearAccessTime() { 8827 bitField0_ = (bitField0_ & ~0x00000008); 8828 accessTime_ = 0L; 8829 onChanged(); 8830 return this; 8831 } 8832 8833 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INodeSymlink) 8834 } 8835 8836 static { 8837 defaultInstance = new INodeSymlink(true); 8838 defaultInstance.initFields(); 8839 } 8840 8841 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INodeSymlink) 8842 } 8843 8844 public interface INodeOrBuilder 8845 extends com.google.protobuf.MessageOrBuilder { 8846 8847 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 8848 /** 8849 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 8850 */ 8851 boolean hasType(); 8852 /** 8853 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 8854 */ 8855 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType(); 8856 8857 // required uint64 id = 2; 8858 /** 8859 * <code>required uint64 id = 2;</code> 8860 */ 8861 boolean hasId(); 8862 /** 8863 * <code>required uint64 id = 2;</code> 8864 */ 8865 long getId(); 8866 8867 // optional bytes name = 3; 8868 /** 8869 * <code>optional bytes name = 3;</code> 8870 */ 8871 boolean hasName(); 8872 /** 8873 * <code>optional bytes name = 3;</code> 8874 */ 8875 com.google.protobuf.ByteString getName(); 8876 8877 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 8878 /** 8879 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8880 */ 8881 boolean hasFile(); 8882 /** 8883 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8884 */ 8885 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile(); 8886 /** 8887 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 8888 */ 8889 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder(); 8890 8891 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 8892 /** 8893 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8894 */ 8895 boolean hasDirectory(); 8896 /** 8897 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8898 */ 8899 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory(); 8900 /** 8901 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 8902 */ 8903 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder(); 8904 8905 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 8906 /** 8907 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8908 */ 8909 boolean hasSymlink(); 8910 /** 8911 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8912 */ 8913 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink(); 8914 /** 8915 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 8916 */ 8917 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder(); 8918 } 8919 /** 8920 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INode} 8921 */ 8922 public static final class INode extends 8923 com.google.protobuf.GeneratedMessage 8924 implements INodeOrBuilder { 8925 // Use INode.newBuilder() to construct. 8926 private INode(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8927 super(builder); 8928 this.unknownFields = builder.getUnknownFields(); 8929 } 8930 private INode(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8931 8932 private static final INode defaultInstance; 8933 public static INode getDefaultInstance() { 8934 return defaultInstance; 8935 } 8936 8937 public INode getDefaultInstanceForType() { 8938 return defaultInstance; 8939 } 8940 8941 private final com.google.protobuf.UnknownFieldSet unknownFields; 8942 @java.lang.Override 8943 public final com.google.protobuf.UnknownFieldSet 8944 getUnknownFields() { 8945 return this.unknownFields; 8946 } 8947 private INode( 8948 com.google.protobuf.CodedInputStream input, 8949 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8950 throws com.google.protobuf.InvalidProtocolBufferException { 8951 initFields(); 8952 int mutable_bitField0_ = 0; 8953 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8954 com.google.protobuf.UnknownFieldSet.newBuilder(); 8955 try { 8956 boolean done = false; 8957 while (!done) { 8958 int tag = input.readTag(); 8959 switch (tag) { 8960 case 0: 8961 done = true; 8962 break; 8963 default: { 8964 if (!parseUnknownField(input, unknownFields, 8965 extensionRegistry, tag)) { 8966 done = true; 8967 } 8968 break; 8969 } 8970 case 8: { 8971 int rawValue = input.readEnum(); 8972 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type value = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.valueOf(rawValue); 8973 if (value == null) { 8974 unknownFields.mergeVarintField(1, rawValue); 8975 } else { 8976 bitField0_ |= 0x00000001; 8977 type_ = value; 8978 } 8979 break; 8980 } 8981 case 16: { 8982 bitField0_ |= 0x00000002; 8983 id_ = input.readUInt64(); 8984 break; 8985 } 8986 case 26: { 8987 bitField0_ |= 0x00000004; 8988 name_ = input.readBytes(); 8989 break; 8990 } 8991 case 34: { 8992 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder subBuilder = null; 8993 if (((bitField0_ & 0x00000008) == 0x00000008)) { 8994 subBuilder = file_.toBuilder(); 8995 } 8996 file_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.PARSER, extensionRegistry); 8997 if (subBuilder != null) { 8998 subBuilder.mergeFrom(file_); 8999 file_ = subBuilder.buildPartial(); 9000 } 9001 bitField0_ |= 0x00000008; 9002 break; 9003 } 9004 case 42: { 9005 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder subBuilder = null; 9006 if (((bitField0_ & 0x00000010) == 0x00000010)) { 9007 subBuilder = directory_.toBuilder(); 9008 } 9009 directory_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.PARSER, extensionRegistry); 9010 if (subBuilder != null) { 9011 subBuilder.mergeFrom(directory_); 9012 directory_ = subBuilder.buildPartial(); 9013 } 9014 bitField0_ |= 0x00000010; 9015 break; 9016 } 9017 case 50: { 9018 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder subBuilder = null; 9019 if (((bitField0_ & 0x00000020) == 0x00000020)) { 9020 subBuilder = symlink_.toBuilder(); 9021 } 9022 symlink_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.PARSER, extensionRegistry); 9023 if (subBuilder != null) { 9024 subBuilder.mergeFrom(symlink_); 9025 symlink_ = subBuilder.buildPartial(); 9026 } 9027 bitField0_ |= 0x00000020; 9028 break; 9029 } 9030 } 9031 } 9032 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9033 throw e.setUnfinishedMessage(this); 9034 } catch (java.io.IOException e) { 9035 throw new com.google.protobuf.InvalidProtocolBufferException( 9036 e.getMessage()).setUnfinishedMessage(this); 9037 } finally { 9038 this.unknownFields = unknownFields.build(); 9039 makeExtensionsImmutable(); 9040 } 9041 } 9042 public static final com.google.protobuf.Descriptors.Descriptor 9043 getDescriptor() { 9044 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 9045 } 9046 9047 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9048 internalGetFieldAccessorTable() { 9049 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable 9050 .ensureFieldAccessorsInitialized( 9051 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder.class); 9052 } 9053 9054 public static com.google.protobuf.Parser<INode> PARSER = 9055 new com.google.protobuf.AbstractParser<INode>() { 9056 public INode parsePartialFrom( 9057 com.google.protobuf.CodedInputStream input, 9058 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9059 throws com.google.protobuf.InvalidProtocolBufferException { 9060 return new INode(input, extensionRegistry); 9061 } 9062 }; 9063 9064 @java.lang.Override 9065 public com.google.protobuf.Parser<INode> getParserForType() { 9066 return PARSER; 9067 } 9068 9069 /** 9070 * Protobuf enum {@code hadoop.hdfs.fsimage.INodeSection.INode.Type} 9071 */ 9072 public enum Type 9073 implements com.google.protobuf.ProtocolMessageEnum { 9074 /** 9075 * <code>FILE = 1;</code> 9076 */ 9077 FILE(0, 1), 9078 /** 9079 * <code>DIRECTORY = 2;</code> 9080 */ 9081 DIRECTORY(1, 2), 9082 /** 9083 * <code>SYMLINK = 3;</code> 9084 */ 9085 SYMLINK(2, 3), 9086 ; 9087 9088 /** 9089 * <code>FILE = 1;</code> 9090 */ 9091 public static final int FILE_VALUE = 1; 9092 /** 9093 * <code>DIRECTORY = 2;</code> 9094 */ 9095 public static final int DIRECTORY_VALUE = 2; 9096 /** 9097 * <code>SYMLINK = 3;</code> 9098 */ 9099 public static final int SYMLINK_VALUE = 3; 9100 9101 9102 public final int getNumber() { return value; } 9103 9104 public static Type valueOf(int value) { 9105 switch (value) { 9106 case 1: return FILE; 9107 case 2: return DIRECTORY; 9108 case 3: return SYMLINK; 9109 default: return null; 9110 } 9111 } 9112 9113 public static com.google.protobuf.Internal.EnumLiteMap<Type> 9114 internalGetValueMap() { 9115 return internalValueMap; 9116 } 9117 private static com.google.protobuf.Internal.EnumLiteMap<Type> 9118 internalValueMap = 9119 new com.google.protobuf.Internal.EnumLiteMap<Type>() { 9120 public Type findValueByNumber(int number) { 9121 return Type.valueOf(number); 9122 } 9123 }; 9124 9125 public final com.google.protobuf.Descriptors.EnumValueDescriptor 9126 getValueDescriptor() { 9127 return getDescriptor().getValues().get(index); 9128 } 9129 public final com.google.protobuf.Descriptors.EnumDescriptor 9130 getDescriptorForType() { 9131 return getDescriptor(); 9132 } 9133 public static final com.google.protobuf.Descriptors.EnumDescriptor 9134 getDescriptor() { 9135 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDescriptor().getEnumTypes().get(0); 9136 } 9137 9138 private static final Type[] VALUES = values(); 9139 9140 public static Type valueOf( 9141 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 9142 if (desc.getType() != getDescriptor()) { 9143 throw new java.lang.IllegalArgumentException( 9144 "EnumValueDescriptor is not for this type."); 9145 } 9146 return VALUES[desc.getIndex()]; 9147 } 9148 9149 private final int index; 9150 private final int value; 9151 9152 private Type(int index, int value) { 9153 this.index = index; 9154 this.value = value; 9155 } 9156 9157 // @@protoc_insertion_point(enum_scope:hadoop.hdfs.fsimage.INodeSection.INode.Type) 9158 } 9159 9160 private int bitField0_; 9161 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 9162 public static final int TYPE_FIELD_NUMBER = 1; 9163 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type type_; 9164 /** 9165 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9166 */ 9167 public boolean hasType() { 9168 return ((bitField0_ & 0x00000001) == 0x00000001); 9169 } 9170 /** 9171 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9172 */ 9173 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType() { 9174 return type_; 9175 } 9176 9177 // required uint64 id = 2; 9178 public static final int ID_FIELD_NUMBER = 2; 9179 private long id_; 9180 /** 9181 * <code>required uint64 id = 2;</code> 9182 */ 9183 public boolean hasId() { 9184 return ((bitField0_ & 0x00000002) == 0x00000002); 9185 } 9186 /** 9187 * <code>required uint64 id = 2;</code> 9188 */ 9189 public long getId() { 9190 return id_; 9191 } 9192 9193 // optional bytes name = 3; 9194 public static final int NAME_FIELD_NUMBER = 3; 9195 private com.google.protobuf.ByteString name_; 9196 /** 9197 * <code>optional bytes name = 3;</code> 9198 */ 9199 public boolean hasName() { 9200 return ((bitField0_ & 0x00000004) == 0x00000004); 9201 } 9202 /** 9203 * <code>optional bytes name = 3;</code> 9204 */ 9205 public com.google.protobuf.ByteString getName() { 9206 return name_; 9207 } 9208 9209 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 9210 public static final int FILE_FIELD_NUMBER = 4; 9211 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile file_; 9212 /** 9213 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9214 */ 9215 public boolean hasFile() { 9216 return ((bitField0_ & 0x00000008) == 0x00000008); 9217 } 9218 /** 9219 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9220 */ 9221 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile() { 9222 return file_; 9223 } 9224 /** 9225 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9226 */ 9227 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder() { 9228 return file_; 9229 } 9230 9231 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 9232 public static final int DIRECTORY_FIELD_NUMBER = 5; 9233 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory directory_; 9234 /** 9235 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9236 */ 9237 public boolean hasDirectory() { 9238 return ((bitField0_ & 0x00000010) == 0x00000010); 9239 } 9240 /** 9241 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9242 */ 9243 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory() { 9244 return directory_; 9245 } 9246 /** 9247 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9248 */ 9249 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder() { 9250 return directory_; 9251 } 9252 9253 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 9254 public static final int SYMLINK_FIELD_NUMBER = 6; 9255 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink symlink_; 9256 /** 9257 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 9258 */ 9259 public boolean hasSymlink() { 9260 return ((bitField0_ & 0x00000020) == 0x00000020); 9261 } 9262 /** 9263 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 9264 */ 9265 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink() { 9266 return symlink_; 9267 } 9268 /** 9269 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 9270 */ 9271 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder() { 9272 return symlink_; 9273 } 9274 9275 private void initFields() { 9276 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 9277 id_ = 0L; 9278 name_ = com.google.protobuf.ByteString.EMPTY; 9279 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 9280 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 9281 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 9282 } 9283 private byte memoizedIsInitialized = -1; 9284 public final boolean isInitialized() { 9285 byte isInitialized = memoizedIsInitialized; 9286 if (isInitialized != -1) return isInitialized == 1; 9287 9288 if (!hasType()) { 9289 memoizedIsInitialized = 0; 9290 return false; 9291 } 9292 if (!hasId()) { 9293 memoizedIsInitialized = 0; 9294 return false; 9295 } 9296 if (hasFile()) { 9297 if (!getFile().isInitialized()) { 9298 memoizedIsInitialized = 0; 9299 return false; 9300 } 9301 } 9302 if (hasDirectory()) { 9303 if (!getDirectory().isInitialized()) { 9304 memoizedIsInitialized = 0; 9305 return false; 9306 } 9307 } 9308 memoizedIsInitialized = 1; 9309 return true; 9310 } 9311 9312 public void writeTo(com.google.protobuf.CodedOutputStream output) 9313 throws java.io.IOException { 9314 getSerializedSize(); 9315 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9316 output.writeEnum(1, type_.getNumber()); 9317 } 9318 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9319 output.writeUInt64(2, id_); 9320 } 9321 if (((bitField0_ & 0x00000004) == 0x00000004)) { 9322 output.writeBytes(3, name_); 9323 } 9324 if (((bitField0_ & 0x00000008) == 0x00000008)) { 9325 output.writeMessage(4, file_); 9326 } 9327 if (((bitField0_ & 0x00000010) == 0x00000010)) { 9328 output.writeMessage(5, directory_); 9329 } 9330 if (((bitField0_ & 0x00000020) == 0x00000020)) { 9331 output.writeMessage(6, symlink_); 9332 } 9333 getUnknownFields().writeTo(output); 9334 } 9335 9336 private int memoizedSerializedSize = -1; 9337 public int getSerializedSize() { 9338 int size = memoizedSerializedSize; 9339 if (size != -1) return size; 9340 9341 size = 0; 9342 if (((bitField0_ & 0x00000001) == 0x00000001)) { 9343 size += com.google.protobuf.CodedOutputStream 9344 .computeEnumSize(1, type_.getNumber()); 9345 } 9346 if (((bitField0_ & 0x00000002) == 0x00000002)) { 9347 size += com.google.protobuf.CodedOutputStream 9348 .computeUInt64Size(2, id_); 9349 } 9350 if (((bitField0_ & 0x00000004) == 0x00000004)) { 9351 size += com.google.protobuf.CodedOutputStream 9352 .computeBytesSize(3, name_); 9353 } 9354 if (((bitField0_ & 0x00000008) == 0x00000008)) { 9355 size += com.google.protobuf.CodedOutputStream 9356 .computeMessageSize(4, file_); 9357 } 9358 if (((bitField0_ & 0x00000010) == 0x00000010)) { 9359 size += com.google.protobuf.CodedOutputStream 9360 .computeMessageSize(5, directory_); 9361 } 9362 if (((bitField0_ & 0x00000020) == 0x00000020)) { 9363 size += com.google.protobuf.CodedOutputStream 9364 .computeMessageSize(6, symlink_); 9365 } 9366 size += getUnknownFields().getSerializedSize(); 9367 memoizedSerializedSize = size; 9368 return size; 9369 } 9370 9371 private static final long serialVersionUID = 0L; 9372 @java.lang.Override 9373 protected java.lang.Object writeReplace() 9374 throws java.io.ObjectStreamException { 9375 return super.writeReplace(); 9376 } 9377 9378 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9379 com.google.protobuf.ByteString data) 9380 throws com.google.protobuf.InvalidProtocolBufferException { 9381 return PARSER.parseFrom(data); 9382 } 9383 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9384 com.google.protobuf.ByteString data, 9385 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9386 throws com.google.protobuf.InvalidProtocolBufferException { 9387 return PARSER.parseFrom(data, extensionRegistry); 9388 } 9389 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom(byte[] data) 9390 throws com.google.protobuf.InvalidProtocolBufferException { 9391 return PARSER.parseFrom(data); 9392 } 9393 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9394 byte[] data, 9395 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9396 throws com.google.protobuf.InvalidProtocolBufferException { 9397 return PARSER.parseFrom(data, extensionRegistry); 9398 } 9399 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom(java.io.InputStream input) 9400 throws java.io.IOException { 9401 return PARSER.parseFrom(input); 9402 } 9403 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9404 java.io.InputStream input, 9405 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9406 throws java.io.IOException { 9407 return PARSER.parseFrom(input, extensionRegistry); 9408 } 9409 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseDelimitedFrom(java.io.InputStream input) 9410 throws java.io.IOException { 9411 return PARSER.parseDelimitedFrom(input); 9412 } 9413 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseDelimitedFrom( 9414 java.io.InputStream input, 9415 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9416 throws java.io.IOException { 9417 return PARSER.parseDelimitedFrom(input, extensionRegistry); 9418 } 9419 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9420 com.google.protobuf.CodedInputStream input) 9421 throws java.io.IOException { 9422 return PARSER.parseFrom(input); 9423 } 9424 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parseFrom( 9425 com.google.protobuf.CodedInputStream input, 9426 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9427 throws java.io.IOException { 9428 return PARSER.parseFrom(input, extensionRegistry); 9429 } 9430 9431 public static Builder newBuilder() { return Builder.create(); } 9432 public Builder newBuilderForType() { return newBuilder(); } 9433 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode prototype) { 9434 return newBuilder().mergeFrom(prototype); 9435 } 9436 public Builder toBuilder() { return newBuilder(this); } 9437 9438 @java.lang.Override 9439 protected Builder newBuilderForType( 9440 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9441 Builder builder = new Builder(parent); 9442 return builder; 9443 } 9444 /** 9445 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection.INode} 9446 */ 9447 public static final class Builder extends 9448 com.google.protobuf.GeneratedMessage.Builder<Builder> 9449 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder { 9450 public static final com.google.protobuf.Descriptors.Descriptor 9451 getDescriptor() { 9452 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 9453 } 9454 9455 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 9456 internalGetFieldAccessorTable() { 9457 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable 9458 .ensureFieldAccessorsInitialized( 9459 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder.class); 9460 } 9461 9462 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.newBuilder() 9463 private Builder() { 9464 maybeForceBuilderInitialization(); 9465 } 9466 9467 private Builder( 9468 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 9469 super(parent); 9470 maybeForceBuilderInitialization(); 9471 } 9472 private void maybeForceBuilderInitialization() { 9473 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 9474 getFileFieldBuilder(); 9475 getDirectoryFieldBuilder(); 9476 getSymlinkFieldBuilder(); 9477 } 9478 } 9479 private static Builder create() { 9480 return new Builder(); 9481 } 9482 9483 public Builder clear() { 9484 super.clear(); 9485 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 9486 bitField0_ = (bitField0_ & ~0x00000001); 9487 id_ = 0L; 9488 bitField0_ = (bitField0_ & ~0x00000002); 9489 name_ = com.google.protobuf.ByteString.EMPTY; 9490 bitField0_ = (bitField0_ & ~0x00000004); 9491 if (fileBuilder_ == null) { 9492 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 9493 } else { 9494 fileBuilder_.clear(); 9495 } 9496 bitField0_ = (bitField0_ & ~0x00000008); 9497 if (directoryBuilder_ == null) { 9498 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 9499 } else { 9500 directoryBuilder_.clear(); 9501 } 9502 bitField0_ = (bitField0_ & ~0x00000010); 9503 if (symlinkBuilder_ == null) { 9504 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 9505 } else { 9506 symlinkBuilder_.clear(); 9507 } 9508 bitField0_ = (bitField0_ & ~0x00000020); 9509 return this; 9510 } 9511 9512 public Builder clone() { 9513 return create().mergeFrom(buildPartial()); 9514 } 9515 9516 public com.google.protobuf.Descriptors.Descriptor 9517 getDescriptorForType() { 9518 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 9519 } 9520 9521 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getDefaultInstanceForType() { 9522 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 9523 } 9524 9525 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode build() { 9526 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode result = buildPartial(); 9527 if (!result.isInitialized()) { 9528 throw newUninitializedMessageException(result); 9529 } 9530 return result; 9531 } 9532 9533 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode buildPartial() { 9534 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode(this); 9535 int from_bitField0_ = bitField0_; 9536 int to_bitField0_ = 0; 9537 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 9538 to_bitField0_ |= 0x00000001; 9539 } 9540 result.type_ = type_; 9541 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 9542 to_bitField0_ |= 0x00000002; 9543 } 9544 result.id_ = id_; 9545 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 9546 to_bitField0_ |= 0x00000004; 9547 } 9548 result.name_ = name_; 9549 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 9550 to_bitField0_ |= 0x00000008; 9551 } 9552 if (fileBuilder_ == null) { 9553 result.file_ = file_; 9554 } else { 9555 result.file_ = fileBuilder_.build(); 9556 } 9557 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 9558 to_bitField0_ |= 0x00000010; 9559 } 9560 if (directoryBuilder_ == null) { 9561 result.directory_ = directory_; 9562 } else { 9563 result.directory_ = directoryBuilder_.build(); 9564 } 9565 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 9566 to_bitField0_ |= 0x00000020; 9567 } 9568 if (symlinkBuilder_ == null) { 9569 result.symlink_ = symlink_; 9570 } else { 9571 result.symlink_ = symlinkBuilder_.build(); 9572 } 9573 result.bitField0_ = to_bitField0_; 9574 onBuilt(); 9575 return result; 9576 } 9577 9578 public Builder mergeFrom(com.google.protobuf.Message other) { 9579 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode) { 9580 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode)other); 9581 } else { 9582 super.mergeFrom(other); 9583 return this; 9584 } 9585 } 9586 9587 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode other) { 9588 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance()) return this; 9589 if (other.hasType()) { 9590 setType(other.getType()); 9591 } 9592 if (other.hasId()) { 9593 setId(other.getId()); 9594 } 9595 if (other.hasName()) { 9596 setName(other.getName()); 9597 } 9598 if (other.hasFile()) { 9599 mergeFile(other.getFile()); 9600 } 9601 if (other.hasDirectory()) { 9602 mergeDirectory(other.getDirectory()); 9603 } 9604 if (other.hasSymlink()) { 9605 mergeSymlink(other.getSymlink()); 9606 } 9607 this.mergeUnknownFields(other.getUnknownFields()); 9608 return this; 9609 } 9610 9611 public final boolean isInitialized() { 9612 if (!hasType()) { 9613 9614 return false; 9615 } 9616 if (!hasId()) { 9617 9618 return false; 9619 } 9620 if (hasFile()) { 9621 if (!getFile().isInitialized()) { 9622 9623 return false; 9624 } 9625 } 9626 if (hasDirectory()) { 9627 if (!getDirectory().isInitialized()) { 9628 9629 return false; 9630 } 9631 } 9632 return true; 9633 } 9634 9635 public Builder mergeFrom( 9636 com.google.protobuf.CodedInputStream input, 9637 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 9638 throws java.io.IOException { 9639 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode parsedMessage = null; 9640 try { 9641 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 9642 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 9643 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode) e.getUnfinishedMessage(); 9644 throw e; 9645 } finally { 9646 if (parsedMessage != null) { 9647 mergeFrom(parsedMessage); 9648 } 9649 } 9650 return this; 9651 } 9652 private int bitField0_; 9653 9654 // required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1; 9655 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 9656 /** 9657 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9658 */ 9659 public boolean hasType() { 9660 return ((bitField0_ & 0x00000001) == 0x00000001); 9661 } 9662 /** 9663 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9664 */ 9665 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type getType() { 9666 return type_; 9667 } 9668 /** 9669 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9670 */ 9671 public Builder setType(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type value) { 9672 if (value == null) { 9673 throw new NullPointerException(); 9674 } 9675 bitField0_ |= 0x00000001; 9676 type_ = value; 9677 onChanged(); 9678 return this; 9679 } 9680 /** 9681 * <code>required .hadoop.hdfs.fsimage.INodeSection.INode.Type type = 1;</code> 9682 */ 9683 public Builder clearType() { 9684 bitField0_ = (bitField0_ & ~0x00000001); 9685 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Type.FILE; 9686 onChanged(); 9687 return this; 9688 } 9689 9690 // required uint64 id = 2; 9691 private long id_ ; 9692 /** 9693 * <code>required uint64 id = 2;</code> 9694 */ 9695 public boolean hasId() { 9696 return ((bitField0_ & 0x00000002) == 0x00000002); 9697 } 9698 /** 9699 * <code>required uint64 id = 2;</code> 9700 */ 9701 public long getId() { 9702 return id_; 9703 } 9704 /** 9705 * <code>required uint64 id = 2;</code> 9706 */ 9707 public Builder setId(long value) { 9708 bitField0_ |= 0x00000002; 9709 id_ = value; 9710 onChanged(); 9711 return this; 9712 } 9713 /** 9714 * <code>required uint64 id = 2;</code> 9715 */ 9716 public Builder clearId() { 9717 bitField0_ = (bitField0_ & ~0x00000002); 9718 id_ = 0L; 9719 onChanged(); 9720 return this; 9721 } 9722 9723 // optional bytes name = 3; 9724 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 9725 /** 9726 * <code>optional bytes name = 3;</code> 9727 */ 9728 public boolean hasName() { 9729 return ((bitField0_ & 0x00000004) == 0x00000004); 9730 } 9731 /** 9732 * <code>optional bytes name = 3;</code> 9733 */ 9734 public com.google.protobuf.ByteString getName() { 9735 return name_; 9736 } 9737 /** 9738 * <code>optional bytes name = 3;</code> 9739 */ 9740 public Builder setName(com.google.protobuf.ByteString value) { 9741 if (value == null) { 9742 throw new NullPointerException(); 9743 } 9744 bitField0_ |= 0x00000004; 9745 name_ = value; 9746 onChanged(); 9747 return this; 9748 } 9749 /** 9750 * <code>optional bytes name = 3;</code> 9751 */ 9752 public Builder clearName() { 9753 bitField0_ = (bitField0_ & ~0x00000004); 9754 name_ = getDefaultInstance().getName(); 9755 onChanged(); 9756 return this; 9757 } 9758 9759 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4; 9760 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 9761 private com.google.protobuf.SingleFieldBuilder< 9762 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> fileBuilder_; 9763 /** 9764 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9765 */ 9766 public boolean hasFile() { 9767 return ((bitField0_ & 0x00000008) == 0x00000008); 9768 } 9769 /** 9770 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9771 */ 9772 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getFile() { 9773 if (fileBuilder_ == null) { 9774 return file_; 9775 } else { 9776 return fileBuilder_.getMessage(); 9777 } 9778 } 9779 /** 9780 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9781 */ 9782 public Builder setFile(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 9783 if (fileBuilder_ == null) { 9784 if (value == null) { 9785 throw new NullPointerException(); 9786 } 9787 file_ = value; 9788 onChanged(); 9789 } else { 9790 fileBuilder_.setMessage(value); 9791 } 9792 bitField0_ |= 0x00000008; 9793 return this; 9794 } 9795 /** 9796 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9797 */ 9798 public Builder setFile( 9799 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder builderForValue) { 9800 if (fileBuilder_ == null) { 9801 file_ = builderForValue.build(); 9802 onChanged(); 9803 } else { 9804 fileBuilder_.setMessage(builderForValue.build()); 9805 } 9806 bitField0_ |= 0x00000008; 9807 return this; 9808 } 9809 /** 9810 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9811 */ 9812 public Builder mergeFile(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 9813 if (fileBuilder_ == null) { 9814 if (((bitField0_ & 0x00000008) == 0x00000008) && 9815 file_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) { 9816 file_ = 9817 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder(file_).mergeFrom(value).buildPartial(); 9818 } else { 9819 file_ = value; 9820 } 9821 onChanged(); 9822 } else { 9823 fileBuilder_.mergeFrom(value); 9824 } 9825 bitField0_ |= 0x00000008; 9826 return this; 9827 } 9828 /** 9829 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9830 */ 9831 public Builder clearFile() { 9832 if (fileBuilder_ == null) { 9833 file_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 9834 onChanged(); 9835 } else { 9836 fileBuilder_.clear(); 9837 } 9838 bitField0_ = (bitField0_ & ~0x00000008); 9839 return this; 9840 } 9841 /** 9842 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9843 */ 9844 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder getFileBuilder() { 9845 bitField0_ |= 0x00000008; 9846 onChanged(); 9847 return getFileFieldBuilder().getBuilder(); 9848 } 9849 /** 9850 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9851 */ 9852 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getFileOrBuilder() { 9853 if (fileBuilder_ != null) { 9854 return fileBuilder_.getMessageOrBuilder(); 9855 } else { 9856 return file_; 9857 } 9858 } 9859 /** 9860 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile file = 4;</code> 9861 */ 9862 private com.google.protobuf.SingleFieldBuilder< 9863 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> 9864 getFileFieldBuilder() { 9865 if (fileBuilder_ == null) { 9866 fileBuilder_ = new com.google.protobuf.SingleFieldBuilder< 9867 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder>( 9868 file_, 9869 getParentForChildren(), 9870 isClean()); 9871 file_ = null; 9872 } 9873 return fileBuilder_; 9874 } 9875 9876 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5; 9877 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 9878 private com.google.protobuf.SingleFieldBuilder< 9879 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> directoryBuilder_; 9880 /** 9881 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9882 */ 9883 public boolean hasDirectory() { 9884 return ((bitField0_ & 0x00000010) == 0x00000010); 9885 } 9886 /** 9887 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9888 */ 9889 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getDirectory() { 9890 if (directoryBuilder_ == null) { 9891 return directory_; 9892 } else { 9893 return directoryBuilder_.getMessage(); 9894 } 9895 } 9896 /** 9897 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9898 */ 9899 public Builder setDirectory(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 9900 if (directoryBuilder_ == null) { 9901 if (value == null) { 9902 throw new NullPointerException(); 9903 } 9904 directory_ = value; 9905 onChanged(); 9906 } else { 9907 directoryBuilder_.setMessage(value); 9908 } 9909 bitField0_ |= 0x00000010; 9910 return this; 9911 } 9912 /** 9913 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9914 */ 9915 public Builder setDirectory( 9916 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder builderForValue) { 9917 if (directoryBuilder_ == null) { 9918 directory_ = builderForValue.build(); 9919 onChanged(); 9920 } else { 9921 directoryBuilder_.setMessage(builderForValue.build()); 9922 } 9923 bitField0_ |= 0x00000010; 9924 return this; 9925 } 9926 /** 9927 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9928 */ 9929 public Builder mergeDirectory(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 9930 if (directoryBuilder_ == null) { 9931 if (((bitField0_ & 0x00000010) == 0x00000010) && 9932 directory_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) { 9933 directory_ = 9934 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder(directory_).mergeFrom(value).buildPartial(); 9935 } else { 9936 directory_ = value; 9937 } 9938 onChanged(); 9939 } else { 9940 directoryBuilder_.mergeFrom(value); 9941 } 9942 bitField0_ |= 0x00000010; 9943 return this; 9944 } 9945 /** 9946 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9947 */ 9948 public Builder clearDirectory() { 9949 if (directoryBuilder_ == null) { 9950 directory_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 9951 onChanged(); 9952 } else { 9953 directoryBuilder_.clear(); 9954 } 9955 bitField0_ = (bitField0_ & ~0x00000010); 9956 return this; 9957 } 9958 /** 9959 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9960 */ 9961 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder getDirectoryBuilder() { 9962 bitField0_ |= 0x00000010; 9963 onChanged(); 9964 return getDirectoryFieldBuilder().getBuilder(); 9965 } 9966 /** 9967 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9968 */ 9969 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getDirectoryOrBuilder() { 9970 if (directoryBuilder_ != null) { 9971 return directoryBuilder_.getMessageOrBuilder(); 9972 } else { 9973 return directory_; 9974 } 9975 } 9976 /** 9977 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory directory = 5;</code> 9978 */ 9979 private com.google.protobuf.SingleFieldBuilder< 9980 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> 9981 getDirectoryFieldBuilder() { 9982 if (directoryBuilder_ == null) { 9983 directoryBuilder_ = new com.google.protobuf.SingleFieldBuilder< 9984 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder>( 9985 directory_, 9986 getParentForChildren(), 9987 isClean()); 9988 directory_ = null; 9989 } 9990 return directoryBuilder_; 9991 } 9992 9993 // optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6; 9994 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 9995 private com.google.protobuf.SingleFieldBuilder< 9996 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder> symlinkBuilder_; 9997 /** 9998 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 9999 */ 10000 public boolean hasSymlink() { 10001 return ((bitField0_ & 0x00000020) == 0x00000020); 10002 } 10003 /** 10004 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10005 */ 10006 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink getSymlink() { 10007 if (symlinkBuilder_ == null) { 10008 return symlink_; 10009 } else { 10010 return symlinkBuilder_.getMessage(); 10011 } 10012 } 10013 /** 10014 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10015 */ 10016 public Builder setSymlink(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink value) { 10017 if (symlinkBuilder_ == null) { 10018 if (value == null) { 10019 throw new NullPointerException(); 10020 } 10021 symlink_ = value; 10022 onChanged(); 10023 } else { 10024 symlinkBuilder_.setMessage(value); 10025 } 10026 bitField0_ |= 0x00000020; 10027 return this; 10028 } 10029 /** 10030 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10031 */ 10032 public Builder setSymlink( 10033 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder builderForValue) { 10034 if (symlinkBuilder_ == null) { 10035 symlink_ = builderForValue.build(); 10036 onChanged(); 10037 } else { 10038 symlinkBuilder_.setMessage(builderForValue.build()); 10039 } 10040 bitField0_ |= 0x00000020; 10041 return this; 10042 } 10043 /** 10044 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10045 */ 10046 public Builder mergeSymlink(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink value) { 10047 if (symlinkBuilder_ == null) { 10048 if (((bitField0_ & 0x00000020) == 0x00000020) && 10049 symlink_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance()) { 10050 symlink_ = 10051 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.newBuilder(symlink_).mergeFrom(value).buildPartial(); 10052 } else { 10053 symlink_ = value; 10054 } 10055 onChanged(); 10056 } else { 10057 symlinkBuilder_.mergeFrom(value); 10058 } 10059 bitField0_ |= 0x00000020; 10060 return this; 10061 } 10062 /** 10063 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10064 */ 10065 public Builder clearSymlink() { 10066 if (symlinkBuilder_ == null) { 10067 symlink_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.getDefaultInstance(); 10068 onChanged(); 10069 } else { 10070 symlinkBuilder_.clear(); 10071 } 10072 bitField0_ = (bitField0_ & ~0x00000020); 10073 return this; 10074 } 10075 /** 10076 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10077 */ 10078 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder getSymlinkBuilder() { 10079 bitField0_ |= 0x00000020; 10080 onChanged(); 10081 return getSymlinkFieldBuilder().getBuilder(); 10082 } 10083 /** 10084 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10085 */ 10086 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder getSymlinkOrBuilder() { 10087 if (symlinkBuilder_ != null) { 10088 return symlinkBuilder_.getMessageOrBuilder(); 10089 } else { 10090 return symlink_; 10091 } 10092 } 10093 /** 10094 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeSymlink symlink = 6;</code> 10095 */ 10096 private com.google.protobuf.SingleFieldBuilder< 10097 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder> 10098 getSymlinkFieldBuilder() { 10099 if (symlinkBuilder_ == null) { 10100 symlinkBuilder_ = new com.google.protobuf.SingleFieldBuilder< 10101 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlink.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeSymlinkOrBuilder>( 10102 symlink_, 10103 getParentForChildren(), 10104 isClean()); 10105 symlink_ = null; 10106 } 10107 return symlinkBuilder_; 10108 } 10109 10110 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection.INode) 10111 } 10112 10113 static { 10114 defaultInstance = new INode(true); 10115 defaultInstance.initFields(); 10116 } 10117 10118 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection.INode) 10119 } 10120 10121 private int bitField0_; 10122 // optional uint64 lastInodeId = 1; 10123 public static final int LASTINODEID_FIELD_NUMBER = 1; 10124 private long lastInodeId_; 10125 /** 10126 * <code>optional uint64 lastInodeId = 1;</code> 10127 */ 10128 public boolean hasLastInodeId() { 10129 return ((bitField0_ & 0x00000001) == 0x00000001); 10130 } 10131 /** 10132 * <code>optional uint64 lastInodeId = 1;</code> 10133 */ 10134 public long getLastInodeId() { 10135 return lastInodeId_; 10136 } 10137 10138 // optional uint64 numInodes = 2; 10139 public static final int NUMINODES_FIELD_NUMBER = 2; 10140 private long numInodes_; 10141 /** 10142 * <code>optional uint64 numInodes = 2;</code> 10143 * 10144 * <pre> 10145 * repeated INodes.. 10146 * </pre> 10147 */ 10148 public boolean hasNumInodes() { 10149 return ((bitField0_ & 0x00000002) == 0x00000002); 10150 } 10151 /** 10152 * <code>optional uint64 numInodes = 2;</code> 10153 * 10154 * <pre> 10155 * repeated INodes.. 10156 * </pre> 10157 */ 10158 public long getNumInodes() { 10159 return numInodes_; 10160 } 10161 10162 private void initFields() { 10163 lastInodeId_ = 0L; 10164 numInodes_ = 0L; 10165 } 10166 private byte memoizedIsInitialized = -1; 10167 public final boolean isInitialized() { 10168 byte isInitialized = memoizedIsInitialized; 10169 if (isInitialized != -1) return isInitialized == 1; 10170 10171 memoizedIsInitialized = 1; 10172 return true; 10173 } 10174 10175 public void writeTo(com.google.protobuf.CodedOutputStream output) 10176 throws java.io.IOException { 10177 getSerializedSize(); 10178 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10179 output.writeUInt64(1, lastInodeId_); 10180 } 10181 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10182 output.writeUInt64(2, numInodes_); 10183 } 10184 getUnknownFields().writeTo(output); 10185 } 10186 10187 private int memoizedSerializedSize = -1; 10188 public int getSerializedSize() { 10189 int size = memoizedSerializedSize; 10190 if (size != -1) return size; 10191 10192 size = 0; 10193 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10194 size += com.google.protobuf.CodedOutputStream 10195 .computeUInt64Size(1, lastInodeId_); 10196 } 10197 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10198 size += com.google.protobuf.CodedOutputStream 10199 .computeUInt64Size(2, numInodes_); 10200 } 10201 size += getUnknownFields().getSerializedSize(); 10202 memoizedSerializedSize = size; 10203 return size; 10204 } 10205 10206 private static final long serialVersionUID = 0L; 10207 @java.lang.Override 10208 protected java.lang.Object writeReplace() 10209 throws java.io.ObjectStreamException { 10210 return super.writeReplace(); 10211 } 10212 10213 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10214 com.google.protobuf.ByteString data) 10215 throws com.google.protobuf.InvalidProtocolBufferException { 10216 return PARSER.parseFrom(data); 10217 } 10218 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10219 com.google.protobuf.ByteString data, 10220 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10221 throws com.google.protobuf.InvalidProtocolBufferException { 10222 return PARSER.parseFrom(data, extensionRegistry); 10223 } 10224 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom(byte[] data) 10225 throws com.google.protobuf.InvalidProtocolBufferException { 10226 return PARSER.parseFrom(data); 10227 } 10228 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10229 byte[] data, 10230 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10231 throws com.google.protobuf.InvalidProtocolBufferException { 10232 return PARSER.parseFrom(data, extensionRegistry); 10233 } 10234 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom(java.io.InputStream input) 10235 throws java.io.IOException { 10236 return PARSER.parseFrom(input); 10237 } 10238 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10239 java.io.InputStream input, 10240 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10241 throws java.io.IOException { 10242 return PARSER.parseFrom(input, extensionRegistry); 10243 } 10244 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseDelimitedFrom(java.io.InputStream input) 10245 throws java.io.IOException { 10246 return PARSER.parseDelimitedFrom(input); 10247 } 10248 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseDelimitedFrom( 10249 java.io.InputStream input, 10250 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10251 throws java.io.IOException { 10252 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10253 } 10254 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10255 com.google.protobuf.CodedInputStream input) 10256 throws java.io.IOException { 10257 return PARSER.parseFrom(input); 10258 } 10259 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parseFrom( 10260 com.google.protobuf.CodedInputStream input, 10261 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10262 throws java.io.IOException { 10263 return PARSER.parseFrom(input, extensionRegistry); 10264 } 10265 10266 public static Builder newBuilder() { return Builder.create(); } 10267 public Builder newBuilderForType() { return newBuilder(); } 10268 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection prototype) { 10269 return newBuilder().mergeFrom(prototype); 10270 } 10271 public Builder toBuilder() { return newBuilder(this); } 10272 10273 @java.lang.Override 10274 protected Builder newBuilderForType( 10275 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10276 Builder builder = new Builder(parent); 10277 return builder; 10278 } 10279 /** 10280 * Protobuf type {@code hadoop.hdfs.fsimage.INodeSection} 10281 * 10282 * <pre> 10283 ** 10284 * Permission is serialized as a 64-bit long. [0:24):[25:48):[48:64) (in Big Endian). 10285 * The first and the second parts are the string ids of the user and 10286 * group name, and the last 16 bits are the permission bits. 10287 * 10288 * Name: INODE 10289 * </pre> 10290 */ 10291 public static final class Builder extends 10292 com.google.protobuf.GeneratedMessage.Builder<Builder> 10293 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSectionOrBuilder { 10294 public static final com.google.protobuf.Descriptors.Descriptor 10295 getDescriptor() { 10296 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 10297 } 10298 10299 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10300 internalGetFieldAccessorTable() { 10301 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable 10302 .ensureFieldAccessorsInitialized( 10303 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.Builder.class); 10304 } 10305 10306 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.newBuilder() 10307 private Builder() { 10308 maybeForceBuilderInitialization(); 10309 } 10310 10311 private Builder( 10312 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10313 super(parent); 10314 maybeForceBuilderInitialization(); 10315 } 10316 private void maybeForceBuilderInitialization() { 10317 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10318 } 10319 } 10320 private static Builder create() { 10321 return new Builder(); 10322 } 10323 10324 public Builder clear() { 10325 super.clear(); 10326 lastInodeId_ = 0L; 10327 bitField0_ = (bitField0_ & ~0x00000001); 10328 numInodes_ = 0L; 10329 bitField0_ = (bitField0_ & ~0x00000002); 10330 return this; 10331 } 10332 10333 public Builder clone() { 10334 return create().mergeFrom(buildPartial()); 10335 } 10336 10337 public com.google.protobuf.Descriptors.Descriptor 10338 getDescriptorForType() { 10339 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 10340 } 10341 10342 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection getDefaultInstanceForType() { 10343 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.getDefaultInstance(); 10344 } 10345 10346 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection build() { 10347 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection result = buildPartial(); 10348 if (!result.isInitialized()) { 10349 throw newUninitializedMessageException(result); 10350 } 10351 return result; 10352 } 10353 10354 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection buildPartial() { 10355 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection(this); 10356 int from_bitField0_ = bitField0_; 10357 int to_bitField0_ = 0; 10358 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 10359 to_bitField0_ |= 0x00000001; 10360 } 10361 result.lastInodeId_ = lastInodeId_; 10362 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 10363 to_bitField0_ |= 0x00000002; 10364 } 10365 result.numInodes_ = numInodes_; 10366 result.bitField0_ = to_bitField0_; 10367 onBuilt(); 10368 return result; 10369 } 10370 10371 public Builder mergeFrom(com.google.protobuf.Message other) { 10372 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection) { 10373 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection)other); 10374 } else { 10375 super.mergeFrom(other); 10376 return this; 10377 } 10378 } 10379 10380 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection other) { 10381 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.getDefaultInstance()) return this; 10382 if (other.hasLastInodeId()) { 10383 setLastInodeId(other.getLastInodeId()); 10384 } 10385 if (other.hasNumInodes()) { 10386 setNumInodes(other.getNumInodes()); 10387 } 10388 this.mergeUnknownFields(other.getUnknownFields()); 10389 return this; 10390 } 10391 10392 public final boolean isInitialized() { 10393 return true; 10394 } 10395 10396 public Builder mergeFrom( 10397 com.google.protobuf.CodedInputStream input, 10398 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10399 throws java.io.IOException { 10400 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection parsedMessage = null; 10401 try { 10402 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 10403 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10404 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection) e.getUnfinishedMessage(); 10405 throw e; 10406 } finally { 10407 if (parsedMessage != null) { 10408 mergeFrom(parsedMessage); 10409 } 10410 } 10411 return this; 10412 } 10413 private int bitField0_; 10414 10415 // optional uint64 lastInodeId = 1; 10416 private long lastInodeId_ ; 10417 /** 10418 * <code>optional uint64 lastInodeId = 1;</code> 10419 */ 10420 public boolean hasLastInodeId() { 10421 return ((bitField0_ & 0x00000001) == 0x00000001); 10422 } 10423 /** 10424 * <code>optional uint64 lastInodeId = 1;</code> 10425 */ 10426 public long getLastInodeId() { 10427 return lastInodeId_; 10428 } 10429 /** 10430 * <code>optional uint64 lastInodeId = 1;</code> 10431 */ 10432 public Builder setLastInodeId(long value) { 10433 bitField0_ |= 0x00000001; 10434 lastInodeId_ = value; 10435 onChanged(); 10436 return this; 10437 } 10438 /** 10439 * <code>optional uint64 lastInodeId = 1;</code> 10440 */ 10441 public Builder clearLastInodeId() { 10442 bitField0_ = (bitField0_ & ~0x00000001); 10443 lastInodeId_ = 0L; 10444 onChanged(); 10445 return this; 10446 } 10447 10448 // optional uint64 numInodes = 2; 10449 private long numInodes_ ; 10450 /** 10451 * <code>optional uint64 numInodes = 2;</code> 10452 * 10453 * <pre> 10454 * repeated INodes.. 10455 * </pre> 10456 */ 10457 public boolean hasNumInodes() { 10458 return ((bitField0_ & 0x00000002) == 0x00000002); 10459 } 10460 /** 10461 * <code>optional uint64 numInodes = 2;</code> 10462 * 10463 * <pre> 10464 * repeated INodes.. 10465 * </pre> 10466 */ 10467 public long getNumInodes() { 10468 return numInodes_; 10469 } 10470 /** 10471 * <code>optional uint64 numInodes = 2;</code> 10472 * 10473 * <pre> 10474 * repeated INodes.. 10475 * </pre> 10476 */ 10477 public Builder setNumInodes(long value) { 10478 bitField0_ |= 0x00000002; 10479 numInodes_ = value; 10480 onChanged(); 10481 return this; 10482 } 10483 /** 10484 * <code>optional uint64 numInodes = 2;</code> 10485 * 10486 * <pre> 10487 * repeated INodes.. 10488 * </pre> 10489 */ 10490 public Builder clearNumInodes() { 10491 bitField0_ = (bitField0_ & ~0x00000002); 10492 numInodes_ = 0L; 10493 onChanged(); 10494 return this; 10495 } 10496 10497 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeSection) 10498 } 10499 10500 static { 10501 defaultInstance = new INodeSection(true); 10502 defaultInstance.initFields(); 10503 } 10504 10505 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeSection) 10506 } 10507 10508 public interface FilesUnderConstructionSectionOrBuilder 10509 extends com.google.protobuf.MessageOrBuilder { 10510 } 10511 /** 10512 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection} 10513 * 10514 * <pre> 10515 ** 10516 * This section records information about under-construction files for 10517 * reconstructing the lease map. 10518 * NAME: FILES_UNDERCONSTRUCTION 10519 * </pre> 10520 */ 10521 public static final class FilesUnderConstructionSection extends 10522 com.google.protobuf.GeneratedMessage 10523 implements FilesUnderConstructionSectionOrBuilder { 10524 // Use FilesUnderConstructionSection.newBuilder() to construct. 10525 private FilesUnderConstructionSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 10526 super(builder); 10527 this.unknownFields = builder.getUnknownFields(); 10528 } 10529 private FilesUnderConstructionSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 10530 10531 private static final FilesUnderConstructionSection defaultInstance; 10532 public static FilesUnderConstructionSection getDefaultInstance() { 10533 return defaultInstance; 10534 } 10535 10536 public FilesUnderConstructionSection getDefaultInstanceForType() { 10537 return defaultInstance; 10538 } 10539 10540 private final com.google.protobuf.UnknownFieldSet unknownFields; 10541 @java.lang.Override 10542 public final com.google.protobuf.UnknownFieldSet 10543 getUnknownFields() { 10544 return this.unknownFields; 10545 } 10546 private FilesUnderConstructionSection( 10547 com.google.protobuf.CodedInputStream input, 10548 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10549 throws com.google.protobuf.InvalidProtocolBufferException { 10550 initFields(); 10551 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 10552 com.google.protobuf.UnknownFieldSet.newBuilder(); 10553 try { 10554 boolean done = false; 10555 while (!done) { 10556 int tag = input.readTag(); 10557 switch (tag) { 10558 case 0: 10559 done = true; 10560 break; 10561 default: { 10562 if (!parseUnknownField(input, unknownFields, 10563 extensionRegistry, tag)) { 10564 done = true; 10565 } 10566 break; 10567 } 10568 } 10569 } 10570 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10571 throw e.setUnfinishedMessage(this); 10572 } catch (java.io.IOException e) { 10573 throw new com.google.protobuf.InvalidProtocolBufferException( 10574 e.getMessage()).setUnfinishedMessage(this); 10575 } finally { 10576 this.unknownFields = unknownFields.build(); 10577 makeExtensionsImmutable(); 10578 } 10579 } 10580 public static final com.google.protobuf.Descriptors.Descriptor 10581 getDescriptor() { 10582 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 10583 } 10584 10585 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10586 internalGetFieldAccessorTable() { 10587 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable 10588 .ensureFieldAccessorsInitialized( 10589 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.Builder.class); 10590 } 10591 10592 public static com.google.protobuf.Parser<FilesUnderConstructionSection> PARSER = 10593 new com.google.protobuf.AbstractParser<FilesUnderConstructionSection>() { 10594 public FilesUnderConstructionSection parsePartialFrom( 10595 com.google.protobuf.CodedInputStream input, 10596 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10597 throws com.google.protobuf.InvalidProtocolBufferException { 10598 return new FilesUnderConstructionSection(input, extensionRegistry); 10599 } 10600 }; 10601 10602 @java.lang.Override 10603 public com.google.protobuf.Parser<FilesUnderConstructionSection> getParserForType() { 10604 return PARSER; 10605 } 10606 10607 public interface FileUnderConstructionEntryOrBuilder 10608 extends com.google.protobuf.MessageOrBuilder { 10609 10610 // optional uint64 inodeId = 1; 10611 /** 10612 * <code>optional uint64 inodeId = 1;</code> 10613 */ 10614 boolean hasInodeId(); 10615 /** 10616 * <code>optional uint64 inodeId = 1;</code> 10617 */ 10618 long getInodeId(); 10619 10620 // optional string fullPath = 2; 10621 /** 10622 * <code>optional string fullPath = 2;</code> 10623 */ 10624 boolean hasFullPath(); 10625 /** 10626 * <code>optional string fullPath = 2;</code> 10627 */ 10628 java.lang.String getFullPath(); 10629 /** 10630 * <code>optional string fullPath = 2;</code> 10631 */ 10632 com.google.protobuf.ByteString 10633 getFullPathBytes(); 10634 } 10635 /** 10636 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry} 10637 */ 10638 public static final class FileUnderConstructionEntry extends 10639 com.google.protobuf.GeneratedMessage 10640 implements FileUnderConstructionEntryOrBuilder { 10641 // Use FileUnderConstructionEntry.newBuilder() to construct. 10642 private FileUnderConstructionEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 10643 super(builder); 10644 this.unknownFields = builder.getUnknownFields(); 10645 } 10646 private FileUnderConstructionEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 10647 10648 private static final FileUnderConstructionEntry defaultInstance; 10649 public static FileUnderConstructionEntry getDefaultInstance() { 10650 return defaultInstance; 10651 } 10652 10653 public FileUnderConstructionEntry getDefaultInstanceForType() { 10654 return defaultInstance; 10655 } 10656 10657 private final com.google.protobuf.UnknownFieldSet unknownFields; 10658 @java.lang.Override 10659 public final com.google.protobuf.UnknownFieldSet 10660 getUnknownFields() { 10661 return this.unknownFields; 10662 } 10663 private FileUnderConstructionEntry( 10664 com.google.protobuf.CodedInputStream input, 10665 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10666 throws com.google.protobuf.InvalidProtocolBufferException { 10667 initFields(); 10668 int mutable_bitField0_ = 0; 10669 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 10670 com.google.protobuf.UnknownFieldSet.newBuilder(); 10671 try { 10672 boolean done = false; 10673 while (!done) { 10674 int tag = input.readTag(); 10675 switch (tag) { 10676 case 0: 10677 done = true; 10678 break; 10679 default: { 10680 if (!parseUnknownField(input, unknownFields, 10681 extensionRegistry, tag)) { 10682 done = true; 10683 } 10684 break; 10685 } 10686 case 8: { 10687 bitField0_ |= 0x00000001; 10688 inodeId_ = input.readUInt64(); 10689 break; 10690 } 10691 case 18: { 10692 bitField0_ |= 0x00000002; 10693 fullPath_ = input.readBytes(); 10694 break; 10695 } 10696 } 10697 } 10698 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 10699 throw e.setUnfinishedMessage(this); 10700 } catch (java.io.IOException e) { 10701 throw new com.google.protobuf.InvalidProtocolBufferException( 10702 e.getMessage()).setUnfinishedMessage(this); 10703 } finally { 10704 this.unknownFields = unknownFields.build(); 10705 makeExtensionsImmutable(); 10706 } 10707 } 10708 public static final com.google.protobuf.Descriptors.Descriptor 10709 getDescriptor() { 10710 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 10711 } 10712 10713 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10714 internalGetFieldAccessorTable() { 10715 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable 10716 .ensureFieldAccessorsInitialized( 10717 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.Builder.class); 10718 } 10719 10720 public static com.google.protobuf.Parser<FileUnderConstructionEntry> PARSER = 10721 new com.google.protobuf.AbstractParser<FileUnderConstructionEntry>() { 10722 public FileUnderConstructionEntry parsePartialFrom( 10723 com.google.protobuf.CodedInputStream input, 10724 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10725 throws com.google.protobuf.InvalidProtocolBufferException { 10726 return new FileUnderConstructionEntry(input, extensionRegistry); 10727 } 10728 }; 10729 10730 @java.lang.Override 10731 public com.google.protobuf.Parser<FileUnderConstructionEntry> getParserForType() { 10732 return PARSER; 10733 } 10734 10735 private int bitField0_; 10736 // optional uint64 inodeId = 1; 10737 public static final int INODEID_FIELD_NUMBER = 1; 10738 private long inodeId_; 10739 /** 10740 * <code>optional uint64 inodeId = 1;</code> 10741 */ 10742 public boolean hasInodeId() { 10743 return ((bitField0_ & 0x00000001) == 0x00000001); 10744 } 10745 /** 10746 * <code>optional uint64 inodeId = 1;</code> 10747 */ 10748 public long getInodeId() { 10749 return inodeId_; 10750 } 10751 10752 // optional string fullPath = 2; 10753 public static final int FULLPATH_FIELD_NUMBER = 2; 10754 private java.lang.Object fullPath_; 10755 /** 10756 * <code>optional string fullPath = 2;</code> 10757 */ 10758 public boolean hasFullPath() { 10759 return ((bitField0_ & 0x00000002) == 0x00000002); 10760 } 10761 /** 10762 * <code>optional string fullPath = 2;</code> 10763 */ 10764 public java.lang.String getFullPath() { 10765 java.lang.Object ref = fullPath_; 10766 if (ref instanceof java.lang.String) { 10767 return (java.lang.String) ref; 10768 } else { 10769 com.google.protobuf.ByteString bs = 10770 (com.google.protobuf.ByteString) ref; 10771 java.lang.String s = bs.toStringUtf8(); 10772 if (bs.isValidUtf8()) { 10773 fullPath_ = s; 10774 } 10775 return s; 10776 } 10777 } 10778 /** 10779 * <code>optional string fullPath = 2;</code> 10780 */ 10781 public com.google.protobuf.ByteString 10782 getFullPathBytes() { 10783 java.lang.Object ref = fullPath_; 10784 if (ref instanceof java.lang.String) { 10785 com.google.protobuf.ByteString b = 10786 com.google.protobuf.ByteString.copyFromUtf8( 10787 (java.lang.String) ref); 10788 fullPath_ = b; 10789 return b; 10790 } else { 10791 return (com.google.protobuf.ByteString) ref; 10792 } 10793 } 10794 10795 private void initFields() { 10796 inodeId_ = 0L; 10797 fullPath_ = ""; 10798 } 10799 private byte memoizedIsInitialized = -1; 10800 public final boolean isInitialized() { 10801 byte isInitialized = memoizedIsInitialized; 10802 if (isInitialized != -1) return isInitialized == 1; 10803 10804 memoizedIsInitialized = 1; 10805 return true; 10806 } 10807 10808 public void writeTo(com.google.protobuf.CodedOutputStream output) 10809 throws java.io.IOException { 10810 getSerializedSize(); 10811 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10812 output.writeUInt64(1, inodeId_); 10813 } 10814 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10815 output.writeBytes(2, getFullPathBytes()); 10816 } 10817 getUnknownFields().writeTo(output); 10818 } 10819 10820 private int memoizedSerializedSize = -1; 10821 public int getSerializedSize() { 10822 int size = memoizedSerializedSize; 10823 if (size != -1) return size; 10824 10825 size = 0; 10826 if (((bitField0_ & 0x00000001) == 0x00000001)) { 10827 size += com.google.protobuf.CodedOutputStream 10828 .computeUInt64Size(1, inodeId_); 10829 } 10830 if (((bitField0_ & 0x00000002) == 0x00000002)) { 10831 size += com.google.protobuf.CodedOutputStream 10832 .computeBytesSize(2, getFullPathBytes()); 10833 } 10834 size += getUnknownFields().getSerializedSize(); 10835 memoizedSerializedSize = size; 10836 return size; 10837 } 10838 10839 private static final long serialVersionUID = 0L; 10840 @java.lang.Override 10841 protected java.lang.Object writeReplace() 10842 throws java.io.ObjectStreamException { 10843 return super.writeReplace(); 10844 } 10845 10846 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10847 com.google.protobuf.ByteString data) 10848 throws com.google.protobuf.InvalidProtocolBufferException { 10849 return PARSER.parseFrom(data); 10850 } 10851 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10852 com.google.protobuf.ByteString data, 10853 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10854 throws com.google.protobuf.InvalidProtocolBufferException { 10855 return PARSER.parseFrom(data, extensionRegistry); 10856 } 10857 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom(byte[] data) 10858 throws com.google.protobuf.InvalidProtocolBufferException { 10859 return PARSER.parseFrom(data); 10860 } 10861 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10862 byte[] data, 10863 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10864 throws com.google.protobuf.InvalidProtocolBufferException { 10865 return PARSER.parseFrom(data, extensionRegistry); 10866 } 10867 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom(java.io.InputStream input) 10868 throws java.io.IOException { 10869 return PARSER.parseFrom(input); 10870 } 10871 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10872 java.io.InputStream input, 10873 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10874 throws java.io.IOException { 10875 return PARSER.parseFrom(input, extensionRegistry); 10876 } 10877 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseDelimitedFrom(java.io.InputStream input) 10878 throws java.io.IOException { 10879 return PARSER.parseDelimitedFrom(input); 10880 } 10881 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseDelimitedFrom( 10882 java.io.InputStream input, 10883 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10884 throws java.io.IOException { 10885 return PARSER.parseDelimitedFrom(input, extensionRegistry); 10886 } 10887 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10888 com.google.protobuf.CodedInputStream input) 10889 throws java.io.IOException { 10890 return PARSER.parseFrom(input); 10891 } 10892 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parseFrom( 10893 com.google.protobuf.CodedInputStream input, 10894 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 10895 throws java.io.IOException { 10896 return PARSER.parseFrom(input, extensionRegistry); 10897 } 10898 10899 public static Builder newBuilder() { return Builder.create(); } 10900 public Builder newBuilderForType() { return newBuilder(); } 10901 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry prototype) { 10902 return newBuilder().mergeFrom(prototype); 10903 } 10904 public Builder toBuilder() { return newBuilder(this); } 10905 10906 @java.lang.Override 10907 protected Builder newBuilderForType( 10908 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10909 Builder builder = new Builder(parent); 10910 return builder; 10911 } 10912 /** 10913 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry} 10914 */ 10915 public static final class Builder extends 10916 com.google.protobuf.GeneratedMessage.Builder<Builder> 10917 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntryOrBuilder { 10918 public static final com.google.protobuf.Descriptors.Descriptor 10919 getDescriptor() { 10920 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 10921 } 10922 10923 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 10924 internalGetFieldAccessorTable() { 10925 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable 10926 .ensureFieldAccessorsInitialized( 10927 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.Builder.class); 10928 } 10929 10930 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.newBuilder() 10931 private Builder() { 10932 maybeForceBuilderInitialization(); 10933 } 10934 10935 private Builder( 10936 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 10937 super(parent); 10938 maybeForceBuilderInitialization(); 10939 } 10940 private void maybeForceBuilderInitialization() { 10941 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 10942 } 10943 } 10944 private static Builder create() { 10945 return new Builder(); 10946 } 10947 10948 public Builder clear() { 10949 super.clear(); 10950 inodeId_ = 0L; 10951 bitField0_ = (bitField0_ & ~0x00000001); 10952 fullPath_ = ""; 10953 bitField0_ = (bitField0_ & ~0x00000002); 10954 return this; 10955 } 10956 10957 public Builder clone() { 10958 return create().mergeFrom(buildPartial()); 10959 } 10960 10961 public com.google.protobuf.Descriptors.Descriptor 10962 getDescriptorForType() { 10963 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 10964 } 10965 10966 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry getDefaultInstanceForType() { 10967 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.getDefaultInstance(); 10968 } 10969 10970 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry build() { 10971 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry result = buildPartial(); 10972 if (!result.isInitialized()) { 10973 throw newUninitializedMessageException(result); 10974 } 10975 return result; 10976 } 10977 10978 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry buildPartial() { 10979 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry(this); 10980 int from_bitField0_ = bitField0_; 10981 int to_bitField0_ = 0; 10982 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 10983 to_bitField0_ |= 0x00000001; 10984 } 10985 result.inodeId_ = inodeId_; 10986 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 10987 to_bitField0_ |= 0x00000002; 10988 } 10989 result.fullPath_ = fullPath_; 10990 result.bitField0_ = to_bitField0_; 10991 onBuilt(); 10992 return result; 10993 } 10994 10995 public Builder mergeFrom(com.google.protobuf.Message other) { 10996 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry) { 10997 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry)other); 10998 } else { 10999 super.mergeFrom(other); 11000 return this; 11001 } 11002 } 11003 11004 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry other) { 11005 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry.getDefaultInstance()) return this; 11006 if (other.hasInodeId()) { 11007 setInodeId(other.getInodeId()); 11008 } 11009 if (other.hasFullPath()) { 11010 bitField0_ |= 0x00000002; 11011 fullPath_ = other.fullPath_; 11012 onChanged(); 11013 } 11014 this.mergeUnknownFields(other.getUnknownFields()); 11015 return this; 11016 } 11017 11018 public final boolean isInitialized() { 11019 return true; 11020 } 11021 11022 public Builder mergeFrom( 11023 com.google.protobuf.CodedInputStream input, 11024 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11025 throws java.io.IOException { 11026 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry parsedMessage = null; 11027 try { 11028 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 11029 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11030 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.FileUnderConstructionEntry) e.getUnfinishedMessage(); 11031 throw e; 11032 } finally { 11033 if (parsedMessage != null) { 11034 mergeFrom(parsedMessage); 11035 } 11036 } 11037 return this; 11038 } 11039 private int bitField0_; 11040 11041 // optional uint64 inodeId = 1; 11042 private long inodeId_ ; 11043 /** 11044 * <code>optional uint64 inodeId = 1;</code> 11045 */ 11046 public boolean hasInodeId() { 11047 return ((bitField0_ & 0x00000001) == 0x00000001); 11048 } 11049 /** 11050 * <code>optional uint64 inodeId = 1;</code> 11051 */ 11052 public long getInodeId() { 11053 return inodeId_; 11054 } 11055 /** 11056 * <code>optional uint64 inodeId = 1;</code> 11057 */ 11058 public Builder setInodeId(long value) { 11059 bitField0_ |= 0x00000001; 11060 inodeId_ = value; 11061 onChanged(); 11062 return this; 11063 } 11064 /** 11065 * <code>optional uint64 inodeId = 1;</code> 11066 */ 11067 public Builder clearInodeId() { 11068 bitField0_ = (bitField0_ & ~0x00000001); 11069 inodeId_ = 0L; 11070 onChanged(); 11071 return this; 11072 } 11073 11074 // optional string fullPath = 2; 11075 private java.lang.Object fullPath_ = ""; 11076 /** 11077 * <code>optional string fullPath = 2;</code> 11078 */ 11079 public boolean hasFullPath() { 11080 return ((bitField0_ & 0x00000002) == 0x00000002); 11081 } 11082 /** 11083 * <code>optional string fullPath = 2;</code> 11084 */ 11085 public java.lang.String getFullPath() { 11086 java.lang.Object ref = fullPath_; 11087 if (!(ref instanceof java.lang.String)) { 11088 java.lang.String s = ((com.google.protobuf.ByteString) ref) 11089 .toStringUtf8(); 11090 fullPath_ = s; 11091 return s; 11092 } else { 11093 return (java.lang.String) ref; 11094 } 11095 } 11096 /** 11097 * <code>optional string fullPath = 2;</code> 11098 */ 11099 public com.google.protobuf.ByteString 11100 getFullPathBytes() { 11101 java.lang.Object ref = fullPath_; 11102 if (ref instanceof String) { 11103 com.google.protobuf.ByteString b = 11104 com.google.protobuf.ByteString.copyFromUtf8( 11105 (java.lang.String) ref); 11106 fullPath_ = b; 11107 return b; 11108 } else { 11109 return (com.google.protobuf.ByteString) ref; 11110 } 11111 } 11112 /** 11113 * <code>optional string fullPath = 2;</code> 11114 */ 11115 public Builder setFullPath( 11116 java.lang.String value) { 11117 if (value == null) { 11118 throw new NullPointerException(); 11119 } 11120 bitField0_ |= 0x00000002; 11121 fullPath_ = value; 11122 onChanged(); 11123 return this; 11124 } 11125 /** 11126 * <code>optional string fullPath = 2;</code> 11127 */ 11128 public Builder clearFullPath() { 11129 bitField0_ = (bitField0_ & ~0x00000002); 11130 fullPath_ = getDefaultInstance().getFullPath(); 11131 onChanged(); 11132 return this; 11133 } 11134 /** 11135 * <code>optional string fullPath = 2;</code> 11136 */ 11137 public Builder setFullPathBytes( 11138 com.google.protobuf.ByteString value) { 11139 if (value == null) { 11140 throw new NullPointerException(); 11141 } 11142 bitField0_ |= 0x00000002; 11143 fullPath_ = value; 11144 onChanged(); 11145 return this; 11146 } 11147 11148 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry) 11149 } 11150 11151 static { 11152 defaultInstance = new FileUnderConstructionEntry(true); 11153 defaultInstance.initFields(); 11154 } 11155 11156 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection.FileUnderConstructionEntry) 11157 } 11158 11159 private void initFields() { 11160 } 11161 private byte memoizedIsInitialized = -1; 11162 public final boolean isInitialized() { 11163 byte isInitialized = memoizedIsInitialized; 11164 if (isInitialized != -1) return isInitialized == 1; 11165 11166 memoizedIsInitialized = 1; 11167 return true; 11168 } 11169 11170 public void writeTo(com.google.protobuf.CodedOutputStream output) 11171 throws java.io.IOException { 11172 getSerializedSize(); 11173 getUnknownFields().writeTo(output); 11174 } 11175 11176 private int memoizedSerializedSize = -1; 11177 public int getSerializedSize() { 11178 int size = memoizedSerializedSize; 11179 if (size != -1) return size; 11180 11181 size = 0; 11182 size += getUnknownFields().getSerializedSize(); 11183 memoizedSerializedSize = size; 11184 return size; 11185 } 11186 11187 private static final long serialVersionUID = 0L; 11188 @java.lang.Override 11189 protected java.lang.Object writeReplace() 11190 throws java.io.ObjectStreamException { 11191 return super.writeReplace(); 11192 } 11193 11194 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11195 com.google.protobuf.ByteString data) 11196 throws com.google.protobuf.InvalidProtocolBufferException { 11197 return PARSER.parseFrom(data); 11198 } 11199 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11200 com.google.protobuf.ByteString data, 11201 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11202 throws com.google.protobuf.InvalidProtocolBufferException { 11203 return PARSER.parseFrom(data, extensionRegistry); 11204 } 11205 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom(byte[] data) 11206 throws com.google.protobuf.InvalidProtocolBufferException { 11207 return PARSER.parseFrom(data); 11208 } 11209 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11210 byte[] data, 11211 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11212 throws com.google.protobuf.InvalidProtocolBufferException { 11213 return PARSER.parseFrom(data, extensionRegistry); 11214 } 11215 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom(java.io.InputStream input) 11216 throws java.io.IOException { 11217 return PARSER.parseFrom(input); 11218 } 11219 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11220 java.io.InputStream input, 11221 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11222 throws java.io.IOException { 11223 return PARSER.parseFrom(input, extensionRegistry); 11224 } 11225 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseDelimitedFrom(java.io.InputStream input) 11226 throws java.io.IOException { 11227 return PARSER.parseDelimitedFrom(input); 11228 } 11229 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseDelimitedFrom( 11230 java.io.InputStream input, 11231 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11232 throws java.io.IOException { 11233 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11234 } 11235 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11236 com.google.protobuf.CodedInputStream input) 11237 throws java.io.IOException { 11238 return PARSER.parseFrom(input); 11239 } 11240 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parseFrom( 11241 com.google.protobuf.CodedInputStream input, 11242 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11243 throws java.io.IOException { 11244 return PARSER.parseFrom(input, extensionRegistry); 11245 } 11246 11247 public static Builder newBuilder() { return Builder.create(); } 11248 public Builder newBuilderForType() { return newBuilder(); } 11249 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection prototype) { 11250 return newBuilder().mergeFrom(prototype); 11251 } 11252 public Builder toBuilder() { return newBuilder(this); } 11253 11254 @java.lang.Override 11255 protected Builder newBuilderForType( 11256 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11257 Builder builder = new Builder(parent); 11258 return builder; 11259 } 11260 /** 11261 * Protobuf type {@code hadoop.hdfs.fsimage.FilesUnderConstructionSection} 11262 * 11263 * <pre> 11264 ** 11265 * This section records information about under-construction files for 11266 * reconstructing the lease map. 11267 * NAME: FILES_UNDERCONSTRUCTION 11268 * </pre> 11269 */ 11270 public static final class Builder extends 11271 com.google.protobuf.GeneratedMessage.Builder<Builder> 11272 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSectionOrBuilder { 11273 public static final com.google.protobuf.Descriptors.Descriptor 11274 getDescriptor() { 11275 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 11276 } 11277 11278 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11279 internalGetFieldAccessorTable() { 11280 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable 11281 .ensureFieldAccessorsInitialized( 11282 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.Builder.class); 11283 } 11284 11285 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.newBuilder() 11286 private Builder() { 11287 maybeForceBuilderInitialization(); 11288 } 11289 11290 private Builder( 11291 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11292 super(parent); 11293 maybeForceBuilderInitialization(); 11294 } 11295 private void maybeForceBuilderInitialization() { 11296 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 11297 } 11298 } 11299 private static Builder create() { 11300 return new Builder(); 11301 } 11302 11303 public Builder clear() { 11304 super.clear(); 11305 return this; 11306 } 11307 11308 public Builder clone() { 11309 return create().mergeFrom(buildPartial()); 11310 } 11311 11312 public com.google.protobuf.Descriptors.Descriptor 11313 getDescriptorForType() { 11314 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 11315 } 11316 11317 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection getDefaultInstanceForType() { 11318 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.getDefaultInstance(); 11319 } 11320 11321 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection build() { 11322 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection result = buildPartial(); 11323 if (!result.isInitialized()) { 11324 throw newUninitializedMessageException(result); 11325 } 11326 return result; 11327 } 11328 11329 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection buildPartial() { 11330 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection(this); 11331 onBuilt(); 11332 return result; 11333 } 11334 11335 public Builder mergeFrom(com.google.protobuf.Message other) { 11336 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection) { 11337 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection)other); 11338 } else { 11339 super.mergeFrom(other); 11340 return this; 11341 } 11342 } 11343 11344 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection other) { 11345 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection.getDefaultInstance()) return this; 11346 this.mergeUnknownFields(other.getUnknownFields()); 11347 return this; 11348 } 11349 11350 public final boolean isInitialized() { 11351 return true; 11352 } 11353 11354 public Builder mergeFrom( 11355 com.google.protobuf.CodedInputStream input, 11356 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11357 throws java.io.IOException { 11358 org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection parsedMessage = null; 11359 try { 11360 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 11361 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11362 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.FilesUnderConstructionSection) e.getUnfinishedMessage(); 11363 throw e; 11364 } finally { 11365 if (parsedMessage != null) { 11366 mergeFrom(parsedMessage); 11367 } 11368 } 11369 return this; 11370 } 11371 11372 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection) 11373 } 11374 11375 static { 11376 defaultInstance = new FilesUnderConstructionSection(true); 11377 defaultInstance.initFields(); 11378 } 11379 11380 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.FilesUnderConstructionSection) 11381 } 11382 11383 public interface INodeDirectorySectionOrBuilder 11384 extends com.google.protobuf.MessageOrBuilder { 11385 } 11386 /** 11387 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection} 11388 * 11389 * <pre> 11390 ** 11391 * This section records the children of each directories 11392 * NAME: INODE_DIR 11393 * </pre> 11394 */ 11395 public static final class INodeDirectorySection extends 11396 com.google.protobuf.GeneratedMessage 11397 implements INodeDirectorySectionOrBuilder { 11398 // Use INodeDirectorySection.newBuilder() to construct. 11399 private INodeDirectorySection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 11400 super(builder); 11401 this.unknownFields = builder.getUnknownFields(); 11402 } 11403 private INodeDirectorySection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 11404 11405 private static final INodeDirectorySection defaultInstance; 11406 public static INodeDirectorySection getDefaultInstance() { 11407 return defaultInstance; 11408 } 11409 11410 public INodeDirectorySection getDefaultInstanceForType() { 11411 return defaultInstance; 11412 } 11413 11414 private final com.google.protobuf.UnknownFieldSet unknownFields; 11415 @java.lang.Override 11416 public final com.google.protobuf.UnknownFieldSet 11417 getUnknownFields() { 11418 return this.unknownFields; 11419 } 11420 private INodeDirectorySection( 11421 com.google.protobuf.CodedInputStream input, 11422 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11423 throws com.google.protobuf.InvalidProtocolBufferException { 11424 initFields(); 11425 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 11426 com.google.protobuf.UnknownFieldSet.newBuilder(); 11427 try { 11428 boolean done = false; 11429 while (!done) { 11430 int tag = input.readTag(); 11431 switch (tag) { 11432 case 0: 11433 done = true; 11434 break; 11435 default: { 11436 if (!parseUnknownField(input, unknownFields, 11437 extensionRegistry, tag)) { 11438 done = true; 11439 } 11440 break; 11441 } 11442 } 11443 } 11444 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11445 throw e.setUnfinishedMessage(this); 11446 } catch (java.io.IOException e) { 11447 throw new com.google.protobuf.InvalidProtocolBufferException( 11448 e.getMessage()).setUnfinishedMessage(this); 11449 } finally { 11450 this.unknownFields = unknownFields.build(); 11451 makeExtensionsImmutable(); 11452 } 11453 } 11454 public static final com.google.protobuf.Descriptors.Descriptor 11455 getDescriptor() { 11456 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 11457 } 11458 11459 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11460 internalGetFieldAccessorTable() { 11461 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable 11462 .ensureFieldAccessorsInitialized( 11463 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.Builder.class); 11464 } 11465 11466 public static com.google.protobuf.Parser<INodeDirectorySection> PARSER = 11467 new com.google.protobuf.AbstractParser<INodeDirectorySection>() { 11468 public INodeDirectorySection parsePartialFrom( 11469 com.google.protobuf.CodedInputStream input, 11470 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11471 throws com.google.protobuf.InvalidProtocolBufferException { 11472 return new INodeDirectorySection(input, extensionRegistry); 11473 } 11474 }; 11475 11476 @java.lang.Override 11477 public com.google.protobuf.Parser<INodeDirectorySection> getParserForType() { 11478 return PARSER; 11479 } 11480 11481 public interface DirEntryOrBuilder 11482 extends com.google.protobuf.MessageOrBuilder { 11483 11484 // optional uint64 parent = 1; 11485 /** 11486 * <code>optional uint64 parent = 1;</code> 11487 */ 11488 boolean hasParent(); 11489 /** 11490 * <code>optional uint64 parent = 1;</code> 11491 */ 11492 long getParent(); 11493 11494 // repeated uint64 children = 2 [packed = true]; 11495 /** 11496 * <code>repeated uint64 children = 2 [packed = true];</code> 11497 * 11498 * <pre> 11499 * children that are not reference nodes 11500 * </pre> 11501 */ 11502 java.util.List<java.lang.Long> getChildrenList(); 11503 /** 11504 * <code>repeated uint64 children = 2 [packed = true];</code> 11505 * 11506 * <pre> 11507 * children that are not reference nodes 11508 * </pre> 11509 */ 11510 int getChildrenCount(); 11511 /** 11512 * <code>repeated uint64 children = 2 [packed = true];</code> 11513 * 11514 * <pre> 11515 * children that are not reference nodes 11516 * </pre> 11517 */ 11518 long getChildren(int index); 11519 11520 // repeated uint32 refChildren = 3 [packed = true]; 11521 /** 11522 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11523 * 11524 * <pre> 11525 * children that are reference nodes, each element is a reference node id 11526 * </pre> 11527 */ 11528 java.util.List<java.lang.Integer> getRefChildrenList(); 11529 /** 11530 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11531 * 11532 * <pre> 11533 * children that are reference nodes, each element is a reference node id 11534 * </pre> 11535 */ 11536 int getRefChildrenCount(); 11537 /** 11538 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11539 * 11540 * <pre> 11541 * children that are reference nodes, each element is a reference node id 11542 * </pre> 11543 */ 11544 int getRefChildren(int index); 11545 } 11546 /** 11547 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry} 11548 * 11549 * <pre> 11550 ** 11551 * A single DirEntry needs to fit in the default PB max message size of 11552 * 64MB. Please be careful when adding more fields to a DirEntry! 11553 * </pre> 11554 */ 11555 public static final class DirEntry extends 11556 com.google.protobuf.GeneratedMessage 11557 implements DirEntryOrBuilder { 11558 // Use DirEntry.newBuilder() to construct. 11559 private DirEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 11560 super(builder); 11561 this.unknownFields = builder.getUnknownFields(); 11562 } 11563 private DirEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 11564 11565 private static final DirEntry defaultInstance; 11566 public static DirEntry getDefaultInstance() { 11567 return defaultInstance; 11568 } 11569 11570 public DirEntry getDefaultInstanceForType() { 11571 return defaultInstance; 11572 } 11573 11574 private final com.google.protobuf.UnknownFieldSet unknownFields; 11575 @java.lang.Override 11576 public final com.google.protobuf.UnknownFieldSet 11577 getUnknownFields() { 11578 return this.unknownFields; 11579 } 11580 private DirEntry( 11581 com.google.protobuf.CodedInputStream input, 11582 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11583 throws com.google.protobuf.InvalidProtocolBufferException { 11584 initFields(); 11585 int mutable_bitField0_ = 0; 11586 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 11587 com.google.protobuf.UnknownFieldSet.newBuilder(); 11588 try { 11589 boolean done = false; 11590 while (!done) { 11591 int tag = input.readTag(); 11592 switch (tag) { 11593 case 0: 11594 done = true; 11595 break; 11596 default: { 11597 if (!parseUnknownField(input, unknownFields, 11598 extensionRegistry, tag)) { 11599 done = true; 11600 } 11601 break; 11602 } 11603 case 8: { 11604 bitField0_ |= 0x00000001; 11605 parent_ = input.readUInt64(); 11606 break; 11607 } 11608 case 16: { 11609 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 11610 children_ = new java.util.ArrayList<java.lang.Long>(); 11611 mutable_bitField0_ |= 0x00000002; 11612 } 11613 children_.add(input.readUInt64()); 11614 break; 11615 } 11616 case 18: { 11617 int length = input.readRawVarint32(); 11618 int limit = input.pushLimit(length); 11619 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) { 11620 children_ = new java.util.ArrayList<java.lang.Long>(); 11621 mutable_bitField0_ |= 0x00000002; 11622 } 11623 while (input.getBytesUntilLimit() > 0) { 11624 children_.add(input.readUInt64()); 11625 } 11626 input.popLimit(limit); 11627 break; 11628 } 11629 case 24: { 11630 if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { 11631 refChildren_ = new java.util.ArrayList<java.lang.Integer>(); 11632 mutable_bitField0_ |= 0x00000004; 11633 } 11634 refChildren_.add(input.readUInt32()); 11635 break; 11636 } 11637 case 26: { 11638 int length = input.readRawVarint32(); 11639 int limit = input.pushLimit(length); 11640 if (!((mutable_bitField0_ & 0x00000004) == 0x00000004) && input.getBytesUntilLimit() > 0) { 11641 refChildren_ = new java.util.ArrayList<java.lang.Integer>(); 11642 mutable_bitField0_ |= 0x00000004; 11643 } 11644 while (input.getBytesUntilLimit() > 0) { 11645 refChildren_.add(input.readUInt32()); 11646 } 11647 input.popLimit(limit); 11648 break; 11649 } 11650 } 11651 } 11652 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 11653 throw e.setUnfinishedMessage(this); 11654 } catch (java.io.IOException e) { 11655 throw new com.google.protobuf.InvalidProtocolBufferException( 11656 e.getMessage()).setUnfinishedMessage(this); 11657 } finally { 11658 if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 11659 children_ = java.util.Collections.unmodifiableList(children_); 11660 } 11661 if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { 11662 refChildren_ = java.util.Collections.unmodifiableList(refChildren_); 11663 } 11664 this.unknownFields = unknownFields.build(); 11665 makeExtensionsImmutable(); 11666 } 11667 } 11668 public static final com.google.protobuf.Descriptors.Descriptor 11669 getDescriptor() { 11670 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 11671 } 11672 11673 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11674 internalGetFieldAccessorTable() { 11675 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable 11676 .ensureFieldAccessorsInitialized( 11677 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.Builder.class); 11678 } 11679 11680 public static com.google.protobuf.Parser<DirEntry> PARSER = 11681 new com.google.protobuf.AbstractParser<DirEntry>() { 11682 public DirEntry parsePartialFrom( 11683 com.google.protobuf.CodedInputStream input, 11684 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11685 throws com.google.protobuf.InvalidProtocolBufferException { 11686 return new DirEntry(input, extensionRegistry); 11687 } 11688 }; 11689 11690 @java.lang.Override 11691 public com.google.protobuf.Parser<DirEntry> getParserForType() { 11692 return PARSER; 11693 } 11694 11695 private int bitField0_; 11696 // optional uint64 parent = 1; 11697 public static final int PARENT_FIELD_NUMBER = 1; 11698 private long parent_; 11699 /** 11700 * <code>optional uint64 parent = 1;</code> 11701 */ 11702 public boolean hasParent() { 11703 return ((bitField0_ & 0x00000001) == 0x00000001); 11704 } 11705 /** 11706 * <code>optional uint64 parent = 1;</code> 11707 */ 11708 public long getParent() { 11709 return parent_; 11710 } 11711 11712 // repeated uint64 children = 2 [packed = true]; 11713 public static final int CHILDREN_FIELD_NUMBER = 2; 11714 private java.util.List<java.lang.Long> children_; 11715 /** 11716 * <code>repeated uint64 children = 2 [packed = true];</code> 11717 * 11718 * <pre> 11719 * children that are not reference nodes 11720 * </pre> 11721 */ 11722 public java.util.List<java.lang.Long> 11723 getChildrenList() { 11724 return children_; 11725 } 11726 /** 11727 * <code>repeated uint64 children = 2 [packed = true];</code> 11728 * 11729 * <pre> 11730 * children that are not reference nodes 11731 * </pre> 11732 */ 11733 public int getChildrenCount() { 11734 return children_.size(); 11735 } 11736 /** 11737 * <code>repeated uint64 children = 2 [packed = true];</code> 11738 * 11739 * <pre> 11740 * children that are not reference nodes 11741 * </pre> 11742 */ 11743 public long getChildren(int index) { 11744 return children_.get(index); 11745 } 11746 private int childrenMemoizedSerializedSize = -1; 11747 11748 // repeated uint32 refChildren = 3 [packed = true]; 11749 public static final int REFCHILDREN_FIELD_NUMBER = 3; 11750 private java.util.List<java.lang.Integer> refChildren_; 11751 /** 11752 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11753 * 11754 * <pre> 11755 * children that are reference nodes, each element is a reference node id 11756 * </pre> 11757 */ 11758 public java.util.List<java.lang.Integer> 11759 getRefChildrenList() { 11760 return refChildren_; 11761 } 11762 /** 11763 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11764 * 11765 * <pre> 11766 * children that are reference nodes, each element is a reference node id 11767 * </pre> 11768 */ 11769 public int getRefChildrenCount() { 11770 return refChildren_.size(); 11771 } 11772 /** 11773 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 11774 * 11775 * <pre> 11776 * children that are reference nodes, each element is a reference node id 11777 * </pre> 11778 */ 11779 public int getRefChildren(int index) { 11780 return refChildren_.get(index); 11781 } 11782 private int refChildrenMemoizedSerializedSize = -1; 11783 11784 private void initFields() { 11785 parent_ = 0L; 11786 children_ = java.util.Collections.emptyList(); 11787 refChildren_ = java.util.Collections.emptyList(); 11788 } 11789 private byte memoizedIsInitialized = -1; 11790 public final boolean isInitialized() { 11791 byte isInitialized = memoizedIsInitialized; 11792 if (isInitialized != -1) return isInitialized == 1; 11793 11794 memoizedIsInitialized = 1; 11795 return true; 11796 } 11797 11798 public void writeTo(com.google.protobuf.CodedOutputStream output) 11799 throws java.io.IOException { 11800 getSerializedSize(); 11801 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11802 output.writeUInt64(1, parent_); 11803 } 11804 if (getChildrenList().size() > 0) { 11805 output.writeRawVarint32(18); 11806 output.writeRawVarint32(childrenMemoizedSerializedSize); 11807 } 11808 for (int i = 0; i < children_.size(); i++) { 11809 output.writeUInt64NoTag(children_.get(i)); 11810 } 11811 if (getRefChildrenList().size() > 0) { 11812 output.writeRawVarint32(26); 11813 output.writeRawVarint32(refChildrenMemoizedSerializedSize); 11814 } 11815 for (int i = 0; i < refChildren_.size(); i++) { 11816 output.writeUInt32NoTag(refChildren_.get(i)); 11817 } 11818 getUnknownFields().writeTo(output); 11819 } 11820 11821 private int memoizedSerializedSize = -1; 11822 public int getSerializedSize() { 11823 int size = memoizedSerializedSize; 11824 if (size != -1) return size; 11825 11826 size = 0; 11827 if (((bitField0_ & 0x00000001) == 0x00000001)) { 11828 size += com.google.protobuf.CodedOutputStream 11829 .computeUInt64Size(1, parent_); 11830 } 11831 { 11832 int dataSize = 0; 11833 for (int i = 0; i < children_.size(); i++) { 11834 dataSize += com.google.protobuf.CodedOutputStream 11835 .computeUInt64SizeNoTag(children_.get(i)); 11836 } 11837 size += dataSize; 11838 if (!getChildrenList().isEmpty()) { 11839 size += 1; 11840 size += com.google.protobuf.CodedOutputStream 11841 .computeInt32SizeNoTag(dataSize); 11842 } 11843 childrenMemoizedSerializedSize = dataSize; 11844 } 11845 { 11846 int dataSize = 0; 11847 for (int i = 0; i < refChildren_.size(); i++) { 11848 dataSize += com.google.protobuf.CodedOutputStream 11849 .computeUInt32SizeNoTag(refChildren_.get(i)); 11850 } 11851 size += dataSize; 11852 if (!getRefChildrenList().isEmpty()) { 11853 size += 1; 11854 size += com.google.protobuf.CodedOutputStream 11855 .computeInt32SizeNoTag(dataSize); 11856 } 11857 refChildrenMemoizedSerializedSize = dataSize; 11858 } 11859 size += getUnknownFields().getSerializedSize(); 11860 memoizedSerializedSize = size; 11861 return size; 11862 } 11863 11864 private static final long serialVersionUID = 0L; 11865 @java.lang.Override 11866 protected java.lang.Object writeReplace() 11867 throws java.io.ObjectStreamException { 11868 return super.writeReplace(); 11869 } 11870 11871 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11872 com.google.protobuf.ByteString data) 11873 throws com.google.protobuf.InvalidProtocolBufferException { 11874 return PARSER.parseFrom(data); 11875 } 11876 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11877 com.google.protobuf.ByteString data, 11878 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11879 throws com.google.protobuf.InvalidProtocolBufferException { 11880 return PARSER.parseFrom(data, extensionRegistry); 11881 } 11882 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom(byte[] data) 11883 throws com.google.protobuf.InvalidProtocolBufferException { 11884 return PARSER.parseFrom(data); 11885 } 11886 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11887 byte[] data, 11888 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11889 throws com.google.protobuf.InvalidProtocolBufferException { 11890 return PARSER.parseFrom(data, extensionRegistry); 11891 } 11892 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom(java.io.InputStream input) 11893 throws java.io.IOException { 11894 return PARSER.parseFrom(input); 11895 } 11896 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11897 java.io.InputStream input, 11898 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11899 throws java.io.IOException { 11900 return PARSER.parseFrom(input, extensionRegistry); 11901 } 11902 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseDelimitedFrom(java.io.InputStream input) 11903 throws java.io.IOException { 11904 return PARSER.parseDelimitedFrom(input); 11905 } 11906 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseDelimitedFrom( 11907 java.io.InputStream input, 11908 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11909 throws java.io.IOException { 11910 return PARSER.parseDelimitedFrom(input, extensionRegistry); 11911 } 11912 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11913 com.google.protobuf.CodedInputStream input) 11914 throws java.io.IOException { 11915 return PARSER.parseFrom(input); 11916 } 11917 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parseFrom( 11918 com.google.protobuf.CodedInputStream input, 11919 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 11920 throws java.io.IOException { 11921 return PARSER.parseFrom(input, extensionRegistry); 11922 } 11923 11924 public static Builder newBuilder() { return Builder.create(); } 11925 public Builder newBuilderForType() { return newBuilder(); } 11926 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry prototype) { 11927 return newBuilder().mergeFrom(prototype); 11928 } 11929 public Builder toBuilder() { return newBuilder(this); } 11930 11931 @java.lang.Override 11932 protected Builder newBuilderForType( 11933 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11934 Builder builder = new Builder(parent); 11935 return builder; 11936 } 11937 /** 11938 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry} 11939 * 11940 * <pre> 11941 ** 11942 * A single DirEntry needs to fit in the default PB max message size of 11943 * 64MB. Please be careful when adding more fields to a DirEntry! 11944 * </pre> 11945 */ 11946 public static final class Builder extends 11947 com.google.protobuf.GeneratedMessage.Builder<Builder> 11948 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntryOrBuilder { 11949 public static final com.google.protobuf.Descriptors.Descriptor 11950 getDescriptor() { 11951 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 11952 } 11953 11954 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 11955 internalGetFieldAccessorTable() { 11956 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable 11957 .ensureFieldAccessorsInitialized( 11958 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.Builder.class); 11959 } 11960 11961 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.newBuilder() 11962 private Builder() { 11963 maybeForceBuilderInitialization(); 11964 } 11965 11966 private Builder( 11967 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 11968 super(parent); 11969 maybeForceBuilderInitialization(); 11970 } 11971 private void maybeForceBuilderInitialization() { 11972 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 11973 } 11974 } 11975 private static Builder create() { 11976 return new Builder(); 11977 } 11978 11979 public Builder clear() { 11980 super.clear(); 11981 parent_ = 0L; 11982 bitField0_ = (bitField0_ & ~0x00000001); 11983 children_ = java.util.Collections.emptyList(); 11984 bitField0_ = (bitField0_ & ~0x00000002); 11985 refChildren_ = java.util.Collections.emptyList(); 11986 bitField0_ = (bitField0_ & ~0x00000004); 11987 return this; 11988 } 11989 11990 public Builder clone() { 11991 return create().mergeFrom(buildPartial()); 11992 } 11993 11994 public com.google.protobuf.Descriptors.Descriptor 11995 getDescriptorForType() { 11996 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 11997 } 11998 11999 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry getDefaultInstanceForType() { 12000 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.getDefaultInstance(); 12001 } 12002 12003 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry build() { 12004 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry result = buildPartial(); 12005 if (!result.isInitialized()) { 12006 throw newUninitializedMessageException(result); 12007 } 12008 return result; 12009 } 12010 12011 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry buildPartial() { 12012 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry(this); 12013 int from_bitField0_ = bitField0_; 12014 int to_bitField0_ = 0; 12015 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 12016 to_bitField0_ |= 0x00000001; 12017 } 12018 result.parent_ = parent_; 12019 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12020 children_ = java.util.Collections.unmodifiableList(children_); 12021 bitField0_ = (bitField0_ & ~0x00000002); 12022 } 12023 result.children_ = children_; 12024 if (((bitField0_ & 0x00000004) == 0x00000004)) { 12025 refChildren_ = java.util.Collections.unmodifiableList(refChildren_); 12026 bitField0_ = (bitField0_ & ~0x00000004); 12027 } 12028 result.refChildren_ = refChildren_; 12029 result.bitField0_ = to_bitField0_; 12030 onBuilt(); 12031 return result; 12032 } 12033 12034 public Builder mergeFrom(com.google.protobuf.Message other) { 12035 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry) { 12036 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry)other); 12037 } else { 12038 super.mergeFrom(other); 12039 return this; 12040 } 12041 } 12042 12043 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry other) { 12044 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry.getDefaultInstance()) return this; 12045 if (other.hasParent()) { 12046 setParent(other.getParent()); 12047 } 12048 if (!other.children_.isEmpty()) { 12049 if (children_.isEmpty()) { 12050 children_ = other.children_; 12051 bitField0_ = (bitField0_ & ~0x00000002); 12052 } else { 12053 ensureChildrenIsMutable(); 12054 children_.addAll(other.children_); 12055 } 12056 onChanged(); 12057 } 12058 if (!other.refChildren_.isEmpty()) { 12059 if (refChildren_.isEmpty()) { 12060 refChildren_ = other.refChildren_; 12061 bitField0_ = (bitField0_ & ~0x00000004); 12062 } else { 12063 ensureRefChildrenIsMutable(); 12064 refChildren_.addAll(other.refChildren_); 12065 } 12066 onChanged(); 12067 } 12068 this.mergeUnknownFields(other.getUnknownFields()); 12069 return this; 12070 } 12071 12072 public final boolean isInitialized() { 12073 return true; 12074 } 12075 12076 public Builder mergeFrom( 12077 com.google.protobuf.CodedInputStream input, 12078 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12079 throws java.io.IOException { 12080 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry parsedMessage = null; 12081 try { 12082 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 12083 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12084 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.DirEntry) e.getUnfinishedMessage(); 12085 throw e; 12086 } finally { 12087 if (parsedMessage != null) { 12088 mergeFrom(parsedMessage); 12089 } 12090 } 12091 return this; 12092 } 12093 private int bitField0_; 12094 12095 // optional uint64 parent = 1; 12096 private long parent_ ; 12097 /** 12098 * <code>optional uint64 parent = 1;</code> 12099 */ 12100 public boolean hasParent() { 12101 return ((bitField0_ & 0x00000001) == 0x00000001); 12102 } 12103 /** 12104 * <code>optional uint64 parent = 1;</code> 12105 */ 12106 public long getParent() { 12107 return parent_; 12108 } 12109 /** 12110 * <code>optional uint64 parent = 1;</code> 12111 */ 12112 public Builder setParent(long value) { 12113 bitField0_ |= 0x00000001; 12114 parent_ = value; 12115 onChanged(); 12116 return this; 12117 } 12118 /** 12119 * <code>optional uint64 parent = 1;</code> 12120 */ 12121 public Builder clearParent() { 12122 bitField0_ = (bitField0_ & ~0x00000001); 12123 parent_ = 0L; 12124 onChanged(); 12125 return this; 12126 } 12127 12128 // repeated uint64 children = 2 [packed = true]; 12129 private java.util.List<java.lang.Long> children_ = java.util.Collections.emptyList(); 12130 private void ensureChildrenIsMutable() { 12131 if (!((bitField0_ & 0x00000002) == 0x00000002)) { 12132 children_ = new java.util.ArrayList<java.lang.Long>(children_); 12133 bitField0_ |= 0x00000002; 12134 } 12135 } 12136 /** 12137 * <code>repeated uint64 children = 2 [packed = true];</code> 12138 * 12139 * <pre> 12140 * children that are not reference nodes 12141 * </pre> 12142 */ 12143 public java.util.List<java.lang.Long> 12144 getChildrenList() { 12145 return java.util.Collections.unmodifiableList(children_); 12146 } 12147 /** 12148 * <code>repeated uint64 children = 2 [packed = true];</code> 12149 * 12150 * <pre> 12151 * children that are not reference nodes 12152 * </pre> 12153 */ 12154 public int getChildrenCount() { 12155 return children_.size(); 12156 } 12157 /** 12158 * <code>repeated uint64 children = 2 [packed = true];</code> 12159 * 12160 * <pre> 12161 * children that are not reference nodes 12162 * </pre> 12163 */ 12164 public long getChildren(int index) { 12165 return children_.get(index); 12166 } 12167 /** 12168 * <code>repeated uint64 children = 2 [packed = true];</code> 12169 * 12170 * <pre> 12171 * children that are not reference nodes 12172 * </pre> 12173 */ 12174 public Builder setChildren( 12175 int index, long value) { 12176 ensureChildrenIsMutable(); 12177 children_.set(index, value); 12178 onChanged(); 12179 return this; 12180 } 12181 /** 12182 * <code>repeated uint64 children = 2 [packed = true];</code> 12183 * 12184 * <pre> 12185 * children that are not reference nodes 12186 * </pre> 12187 */ 12188 public Builder addChildren(long value) { 12189 ensureChildrenIsMutable(); 12190 children_.add(value); 12191 onChanged(); 12192 return this; 12193 } 12194 /** 12195 * <code>repeated uint64 children = 2 [packed = true];</code> 12196 * 12197 * <pre> 12198 * children that are not reference nodes 12199 * </pre> 12200 */ 12201 public Builder addAllChildren( 12202 java.lang.Iterable<? extends java.lang.Long> values) { 12203 ensureChildrenIsMutable(); 12204 super.addAll(values, children_); 12205 onChanged(); 12206 return this; 12207 } 12208 /** 12209 * <code>repeated uint64 children = 2 [packed = true];</code> 12210 * 12211 * <pre> 12212 * children that are not reference nodes 12213 * </pre> 12214 */ 12215 public Builder clearChildren() { 12216 children_ = java.util.Collections.emptyList(); 12217 bitField0_ = (bitField0_ & ~0x00000002); 12218 onChanged(); 12219 return this; 12220 } 12221 12222 // repeated uint32 refChildren = 3 [packed = true]; 12223 private java.util.List<java.lang.Integer> refChildren_ = java.util.Collections.emptyList(); 12224 private void ensureRefChildrenIsMutable() { 12225 if (!((bitField0_ & 0x00000004) == 0x00000004)) { 12226 refChildren_ = new java.util.ArrayList<java.lang.Integer>(refChildren_); 12227 bitField0_ |= 0x00000004; 12228 } 12229 } 12230 /** 12231 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12232 * 12233 * <pre> 12234 * children that are reference nodes, each element is a reference node id 12235 * </pre> 12236 */ 12237 public java.util.List<java.lang.Integer> 12238 getRefChildrenList() { 12239 return java.util.Collections.unmodifiableList(refChildren_); 12240 } 12241 /** 12242 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12243 * 12244 * <pre> 12245 * children that are reference nodes, each element is a reference node id 12246 * </pre> 12247 */ 12248 public int getRefChildrenCount() { 12249 return refChildren_.size(); 12250 } 12251 /** 12252 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12253 * 12254 * <pre> 12255 * children that are reference nodes, each element is a reference node id 12256 * </pre> 12257 */ 12258 public int getRefChildren(int index) { 12259 return refChildren_.get(index); 12260 } 12261 /** 12262 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12263 * 12264 * <pre> 12265 * children that are reference nodes, each element is a reference node id 12266 * </pre> 12267 */ 12268 public Builder setRefChildren( 12269 int index, int value) { 12270 ensureRefChildrenIsMutable(); 12271 refChildren_.set(index, value); 12272 onChanged(); 12273 return this; 12274 } 12275 /** 12276 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12277 * 12278 * <pre> 12279 * children that are reference nodes, each element is a reference node id 12280 * </pre> 12281 */ 12282 public Builder addRefChildren(int value) { 12283 ensureRefChildrenIsMutable(); 12284 refChildren_.add(value); 12285 onChanged(); 12286 return this; 12287 } 12288 /** 12289 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12290 * 12291 * <pre> 12292 * children that are reference nodes, each element is a reference node id 12293 * </pre> 12294 */ 12295 public Builder addAllRefChildren( 12296 java.lang.Iterable<? extends java.lang.Integer> values) { 12297 ensureRefChildrenIsMutable(); 12298 super.addAll(values, refChildren_); 12299 onChanged(); 12300 return this; 12301 } 12302 /** 12303 * <code>repeated uint32 refChildren = 3 [packed = true];</code> 12304 * 12305 * <pre> 12306 * children that are reference nodes, each element is a reference node id 12307 * </pre> 12308 */ 12309 public Builder clearRefChildren() { 12310 refChildren_ = java.util.Collections.emptyList(); 12311 bitField0_ = (bitField0_ & ~0x00000004); 12312 onChanged(); 12313 return this; 12314 } 12315 12316 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry) 12317 } 12318 12319 static { 12320 defaultInstance = new DirEntry(true); 12321 defaultInstance.initFields(); 12322 } 12323 12324 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeDirectorySection.DirEntry) 12325 } 12326 12327 private void initFields() { 12328 } 12329 private byte memoizedIsInitialized = -1; 12330 public final boolean isInitialized() { 12331 byte isInitialized = memoizedIsInitialized; 12332 if (isInitialized != -1) return isInitialized == 1; 12333 12334 memoizedIsInitialized = 1; 12335 return true; 12336 } 12337 12338 public void writeTo(com.google.protobuf.CodedOutputStream output) 12339 throws java.io.IOException { 12340 getSerializedSize(); 12341 getUnknownFields().writeTo(output); 12342 } 12343 12344 private int memoizedSerializedSize = -1; 12345 public int getSerializedSize() { 12346 int size = memoizedSerializedSize; 12347 if (size != -1) return size; 12348 12349 size = 0; 12350 size += getUnknownFields().getSerializedSize(); 12351 memoizedSerializedSize = size; 12352 return size; 12353 } 12354 12355 private static final long serialVersionUID = 0L; 12356 @java.lang.Override 12357 protected java.lang.Object writeReplace() 12358 throws java.io.ObjectStreamException { 12359 return super.writeReplace(); 12360 } 12361 12362 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12363 com.google.protobuf.ByteString data) 12364 throws com.google.protobuf.InvalidProtocolBufferException { 12365 return PARSER.parseFrom(data); 12366 } 12367 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12368 com.google.protobuf.ByteString data, 12369 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12370 throws com.google.protobuf.InvalidProtocolBufferException { 12371 return PARSER.parseFrom(data, extensionRegistry); 12372 } 12373 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom(byte[] data) 12374 throws com.google.protobuf.InvalidProtocolBufferException { 12375 return PARSER.parseFrom(data); 12376 } 12377 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12378 byte[] data, 12379 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12380 throws com.google.protobuf.InvalidProtocolBufferException { 12381 return PARSER.parseFrom(data, extensionRegistry); 12382 } 12383 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom(java.io.InputStream input) 12384 throws java.io.IOException { 12385 return PARSER.parseFrom(input); 12386 } 12387 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12388 java.io.InputStream input, 12389 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12390 throws java.io.IOException { 12391 return PARSER.parseFrom(input, extensionRegistry); 12392 } 12393 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseDelimitedFrom(java.io.InputStream input) 12394 throws java.io.IOException { 12395 return PARSER.parseDelimitedFrom(input); 12396 } 12397 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseDelimitedFrom( 12398 java.io.InputStream input, 12399 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12400 throws java.io.IOException { 12401 return PARSER.parseDelimitedFrom(input, extensionRegistry); 12402 } 12403 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12404 com.google.protobuf.CodedInputStream input) 12405 throws java.io.IOException { 12406 return PARSER.parseFrom(input); 12407 } 12408 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parseFrom( 12409 com.google.protobuf.CodedInputStream input, 12410 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12411 throws java.io.IOException { 12412 return PARSER.parseFrom(input, extensionRegistry); 12413 } 12414 12415 public static Builder newBuilder() { return Builder.create(); } 12416 public Builder newBuilderForType() { return newBuilder(); } 12417 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection prototype) { 12418 return newBuilder().mergeFrom(prototype); 12419 } 12420 public Builder toBuilder() { return newBuilder(this); } 12421 12422 @java.lang.Override 12423 protected Builder newBuilderForType( 12424 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12425 Builder builder = new Builder(parent); 12426 return builder; 12427 } 12428 /** 12429 * Protobuf type {@code hadoop.hdfs.fsimage.INodeDirectorySection} 12430 * 12431 * <pre> 12432 ** 12433 * This section records the children of each directories 12434 * NAME: INODE_DIR 12435 * </pre> 12436 */ 12437 public static final class Builder extends 12438 com.google.protobuf.GeneratedMessage.Builder<Builder> 12439 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySectionOrBuilder { 12440 public static final com.google.protobuf.Descriptors.Descriptor 12441 getDescriptor() { 12442 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 12443 } 12444 12445 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12446 internalGetFieldAccessorTable() { 12447 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable 12448 .ensureFieldAccessorsInitialized( 12449 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.Builder.class); 12450 } 12451 12452 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.newBuilder() 12453 private Builder() { 12454 maybeForceBuilderInitialization(); 12455 } 12456 12457 private Builder( 12458 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 12459 super(parent); 12460 maybeForceBuilderInitialization(); 12461 } 12462 private void maybeForceBuilderInitialization() { 12463 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 12464 } 12465 } 12466 private static Builder create() { 12467 return new Builder(); 12468 } 12469 12470 public Builder clear() { 12471 super.clear(); 12472 return this; 12473 } 12474 12475 public Builder clone() { 12476 return create().mergeFrom(buildPartial()); 12477 } 12478 12479 public com.google.protobuf.Descriptors.Descriptor 12480 getDescriptorForType() { 12481 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 12482 } 12483 12484 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection getDefaultInstanceForType() { 12485 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.getDefaultInstance(); 12486 } 12487 12488 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection build() { 12489 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection result = buildPartial(); 12490 if (!result.isInitialized()) { 12491 throw newUninitializedMessageException(result); 12492 } 12493 return result; 12494 } 12495 12496 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection buildPartial() { 12497 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection(this); 12498 onBuilt(); 12499 return result; 12500 } 12501 12502 public Builder mergeFrom(com.google.protobuf.Message other) { 12503 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection) { 12504 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection)other); 12505 } else { 12506 super.mergeFrom(other); 12507 return this; 12508 } 12509 } 12510 12511 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection other) { 12512 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection.getDefaultInstance()) return this; 12513 this.mergeUnknownFields(other.getUnknownFields()); 12514 return this; 12515 } 12516 12517 public final boolean isInitialized() { 12518 return true; 12519 } 12520 12521 public Builder mergeFrom( 12522 com.google.protobuf.CodedInputStream input, 12523 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12524 throws java.io.IOException { 12525 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection parsedMessage = null; 12526 try { 12527 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 12528 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12529 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeDirectorySection) e.getUnfinishedMessage(); 12530 throw e; 12531 } finally { 12532 if (parsedMessage != null) { 12533 mergeFrom(parsedMessage); 12534 } 12535 } 12536 return this; 12537 } 12538 12539 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeDirectorySection) 12540 } 12541 12542 static { 12543 defaultInstance = new INodeDirectorySection(true); 12544 defaultInstance.initFields(); 12545 } 12546 12547 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeDirectorySection) 12548 } 12549 12550 public interface INodeReferenceSectionOrBuilder 12551 extends com.google.protobuf.MessageOrBuilder { 12552 } 12553 /** 12554 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection} 12555 */ 12556 public static final class INodeReferenceSection extends 12557 com.google.protobuf.GeneratedMessage 12558 implements INodeReferenceSectionOrBuilder { 12559 // Use INodeReferenceSection.newBuilder() to construct. 12560 private INodeReferenceSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 12561 super(builder); 12562 this.unknownFields = builder.getUnknownFields(); 12563 } 12564 private INodeReferenceSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 12565 12566 private static final INodeReferenceSection defaultInstance; 12567 public static INodeReferenceSection getDefaultInstance() { 12568 return defaultInstance; 12569 } 12570 12571 public INodeReferenceSection getDefaultInstanceForType() { 12572 return defaultInstance; 12573 } 12574 12575 private final com.google.protobuf.UnknownFieldSet unknownFields; 12576 @java.lang.Override 12577 public final com.google.protobuf.UnknownFieldSet 12578 getUnknownFields() { 12579 return this.unknownFields; 12580 } 12581 private INodeReferenceSection( 12582 com.google.protobuf.CodedInputStream input, 12583 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12584 throws com.google.protobuf.InvalidProtocolBufferException { 12585 initFields(); 12586 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 12587 com.google.protobuf.UnknownFieldSet.newBuilder(); 12588 try { 12589 boolean done = false; 12590 while (!done) { 12591 int tag = input.readTag(); 12592 switch (tag) { 12593 case 0: 12594 done = true; 12595 break; 12596 default: { 12597 if (!parseUnknownField(input, unknownFields, 12598 extensionRegistry, tag)) { 12599 done = true; 12600 } 12601 break; 12602 } 12603 } 12604 } 12605 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12606 throw e.setUnfinishedMessage(this); 12607 } catch (java.io.IOException e) { 12608 throw new com.google.protobuf.InvalidProtocolBufferException( 12609 e.getMessage()).setUnfinishedMessage(this); 12610 } finally { 12611 this.unknownFields = unknownFields.build(); 12612 makeExtensionsImmutable(); 12613 } 12614 } 12615 public static final com.google.protobuf.Descriptors.Descriptor 12616 getDescriptor() { 12617 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 12618 } 12619 12620 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12621 internalGetFieldAccessorTable() { 12622 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable 12623 .ensureFieldAccessorsInitialized( 12624 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.Builder.class); 12625 } 12626 12627 public static com.google.protobuf.Parser<INodeReferenceSection> PARSER = 12628 new com.google.protobuf.AbstractParser<INodeReferenceSection>() { 12629 public INodeReferenceSection parsePartialFrom( 12630 com.google.protobuf.CodedInputStream input, 12631 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12632 throws com.google.protobuf.InvalidProtocolBufferException { 12633 return new INodeReferenceSection(input, extensionRegistry); 12634 } 12635 }; 12636 12637 @java.lang.Override 12638 public com.google.protobuf.Parser<INodeReferenceSection> getParserForType() { 12639 return PARSER; 12640 } 12641 12642 public interface INodeReferenceOrBuilder 12643 extends com.google.protobuf.MessageOrBuilder { 12644 12645 // optional uint64 referredId = 1; 12646 /** 12647 * <code>optional uint64 referredId = 1;</code> 12648 * 12649 * <pre> 12650 * id of the referred inode 12651 * </pre> 12652 */ 12653 boolean hasReferredId(); 12654 /** 12655 * <code>optional uint64 referredId = 1;</code> 12656 * 12657 * <pre> 12658 * id of the referred inode 12659 * </pre> 12660 */ 12661 long getReferredId(); 12662 12663 // optional bytes name = 2; 12664 /** 12665 * <code>optional bytes name = 2;</code> 12666 * 12667 * <pre> 12668 * local name recorded in WithName 12669 * </pre> 12670 */ 12671 boolean hasName(); 12672 /** 12673 * <code>optional bytes name = 2;</code> 12674 * 12675 * <pre> 12676 * local name recorded in WithName 12677 * </pre> 12678 */ 12679 com.google.protobuf.ByteString getName(); 12680 12681 // optional uint32 dstSnapshotId = 3; 12682 /** 12683 * <code>optional uint32 dstSnapshotId = 3;</code> 12684 * 12685 * <pre> 12686 * recorded in DstReference 12687 * </pre> 12688 */ 12689 boolean hasDstSnapshotId(); 12690 /** 12691 * <code>optional uint32 dstSnapshotId = 3;</code> 12692 * 12693 * <pre> 12694 * recorded in DstReference 12695 * </pre> 12696 */ 12697 int getDstSnapshotId(); 12698 12699 // optional uint32 lastSnapshotId = 4; 12700 /** 12701 * <code>optional uint32 lastSnapshotId = 4;</code> 12702 * 12703 * <pre> 12704 * recorded in WithName 12705 * </pre> 12706 */ 12707 boolean hasLastSnapshotId(); 12708 /** 12709 * <code>optional uint32 lastSnapshotId = 4;</code> 12710 * 12711 * <pre> 12712 * recorded in WithName 12713 * </pre> 12714 */ 12715 int getLastSnapshotId(); 12716 } 12717 /** 12718 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference} 12719 */ 12720 public static final class INodeReference extends 12721 com.google.protobuf.GeneratedMessage 12722 implements INodeReferenceOrBuilder { 12723 // Use INodeReference.newBuilder() to construct. 12724 private INodeReference(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 12725 super(builder); 12726 this.unknownFields = builder.getUnknownFields(); 12727 } 12728 private INodeReference(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 12729 12730 private static final INodeReference defaultInstance; 12731 public static INodeReference getDefaultInstance() { 12732 return defaultInstance; 12733 } 12734 12735 public INodeReference getDefaultInstanceForType() { 12736 return defaultInstance; 12737 } 12738 12739 private final com.google.protobuf.UnknownFieldSet unknownFields; 12740 @java.lang.Override 12741 public final com.google.protobuf.UnknownFieldSet 12742 getUnknownFields() { 12743 return this.unknownFields; 12744 } 12745 private INodeReference( 12746 com.google.protobuf.CodedInputStream input, 12747 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12748 throws com.google.protobuf.InvalidProtocolBufferException { 12749 initFields(); 12750 int mutable_bitField0_ = 0; 12751 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 12752 com.google.protobuf.UnknownFieldSet.newBuilder(); 12753 try { 12754 boolean done = false; 12755 while (!done) { 12756 int tag = input.readTag(); 12757 switch (tag) { 12758 case 0: 12759 done = true; 12760 break; 12761 default: { 12762 if (!parseUnknownField(input, unknownFields, 12763 extensionRegistry, tag)) { 12764 done = true; 12765 } 12766 break; 12767 } 12768 case 8: { 12769 bitField0_ |= 0x00000001; 12770 referredId_ = input.readUInt64(); 12771 break; 12772 } 12773 case 18: { 12774 bitField0_ |= 0x00000002; 12775 name_ = input.readBytes(); 12776 break; 12777 } 12778 case 24: { 12779 bitField0_ |= 0x00000004; 12780 dstSnapshotId_ = input.readUInt32(); 12781 break; 12782 } 12783 case 32: { 12784 bitField0_ |= 0x00000008; 12785 lastSnapshotId_ = input.readUInt32(); 12786 break; 12787 } 12788 } 12789 } 12790 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 12791 throw e.setUnfinishedMessage(this); 12792 } catch (java.io.IOException e) { 12793 throw new com.google.protobuf.InvalidProtocolBufferException( 12794 e.getMessage()).setUnfinishedMessage(this); 12795 } finally { 12796 this.unknownFields = unknownFields.build(); 12797 makeExtensionsImmutable(); 12798 } 12799 } 12800 public static final com.google.protobuf.Descriptors.Descriptor 12801 getDescriptor() { 12802 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 12803 } 12804 12805 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 12806 internalGetFieldAccessorTable() { 12807 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable 12808 .ensureFieldAccessorsInitialized( 12809 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.Builder.class); 12810 } 12811 12812 public static com.google.protobuf.Parser<INodeReference> PARSER = 12813 new com.google.protobuf.AbstractParser<INodeReference>() { 12814 public INodeReference parsePartialFrom( 12815 com.google.protobuf.CodedInputStream input, 12816 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12817 throws com.google.protobuf.InvalidProtocolBufferException { 12818 return new INodeReference(input, extensionRegistry); 12819 } 12820 }; 12821 12822 @java.lang.Override 12823 public com.google.protobuf.Parser<INodeReference> getParserForType() { 12824 return PARSER; 12825 } 12826 12827 private int bitField0_; 12828 // optional uint64 referredId = 1; 12829 public static final int REFERREDID_FIELD_NUMBER = 1; 12830 private long referredId_; 12831 /** 12832 * <code>optional uint64 referredId = 1;</code> 12833 * 12834 * <pre> 12835 * id of the referred inode 12836 * </pre> 12837 */ 12838 public boolean hasReferredId() { 12839 return ((bitField0_ & 0x00000001) == 0x00000001); 12840 } 12841 /** 12842 * <code>optional uint64 referredId = 1;</code> 12843 * 12844 * <pre> 12845 * id of the referred inode 12846 * </pre> 12847 */ 12848 public long getReferredId() { 12849 return referredId_; 12850 } 12851 12852 // optional bytes name = 2; 12853 public static final int NAME_FIELD_NUMBER = 2; 12854 private com.google.protobuf.ByteString name_; 12855 /** 12856 * <code>optional bytes name = 2;</code> 12857 * 12858 * <pre> 12859 * local name recorded in WithName 12860 * </pre> 12861 */ 12862 public boolean hasName() { 12863 return ((bitField0_ & 0x00000002) == 0x00000002); 12864 } 12865 /** 12866 * <code>optional bytes name = 2;</code> 12867 * 12868 * <pre> 12869 * local name recorded in WithName 12870 * </pre> 12871 */ 12872 public com.google.protobuf.ByteString getName() { 12873 return name_; 12874 } 12875 12876 // optional uint32 dstSnapshotId = 3; 12877 public static final int DSTSNAPSHOTID_FIELD_NUMBER = 3; 12878 private int dstSnapshotId_; 12879 /** 12880 * <code>optional uint32 dstSnapshotId = 3;</code> 12881 * 12882 * <pre> 12883 * recorded in DstReference 12884 * </pre> 12885 */ 12886 public boolean hasDstSnapshotId() { 12887 return ((bitField0_ & 0x00000004) == 0x00000004); 12888 } 12889 /** 12890 * <code>optional uint32 dstSnapshotId = 3;</code> 12891 * 12892 * <pre> 12893 * recorded in DstReference 12894 * </pre> 12895 */ 12896 public int getDstSnapshotId() { 12897 return dstSnapshotId_; 12898 } 12899 12900 // optional uint32 lastSnapshotId = 4; 12901 public static final int LASTSNAPSHOTID_FIELD_NUMBER = 4; 12902 private int lastSnapshotId_; 12903 /** 12904 * <code>optional uint32 lastSnapshotId = 4;</code> 12905 * 12906 * <pre> 12907 * recorded in WithName 12908 * </pre> 12909 */ 12910 public boolean hasLastSnapshotId() { 12911 return ((bitField0_ & 0x00000008) == 0x00000008); 12912 } 12913 /** 12914 * <code>optional uint32 lastSnapshotId = 4;</code> 12915 * 12916 * <pre> 12917 * recorded in WithName 12918 * </pre> 12919 */ 12920 public int getLastSnapshotId() { 12921 return lastSnapshotId_; 12922 } 12923 12924 private void initFields() { 12925 referredId_ = 0L; 12926 name_ = com.google.protobuf.ByteString.EMPTY; 12927 dstSnapshotId_ = 0; 12928 lastSnapshotId_ = 0; 12929 } 12930 private byte memoizedIsInitialized = -1; 12931 public final boolean isInitialized() { 12932 byte isInitialized = memoizedIsInitialized; 12933 if (isInitialized != -1) return isInitialized == 1; 12934 12935 memoizedIsInitialized = 1; 12936 return true; 12937 } 12938 12939 public void writeTo(com.google.protobuf.CodedOutputStream output) 12940 throws java.io.IOException { 12941 getSerializedSize(); 12942 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12943 output.writeUInt64(1, referredId_); 12944 } 12945 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12946 output.writeBytes(2, name_); 12947 } 12948 if (((bitField0_ & 0x00000004) == 0x00000004)) { 12949 output.writeUInt32(3, dstSnapshotId_); 12950 } 12951 if (((bitField0_ & 0x00000008) == 0x00000008)) { 12952 output.writeUInt32(4, lastSnapshotId_); 12953 } 12954 getUnknownFields().writeTo(output); 12955 } 12956 12957 private int memoizedSerializedSize = -1; 12958 public int getSerializedSize() { 12959 int size = memoizedSerializedSize; 12960 if (size != -1) return size; 12961 12962 size = 0; 12963 if (((bitField0_ & 0x00000001) == 0x00000001)) { 12964 size += com.google.protobuf.CodedOutputStream 12965 .computeUInt64Size(1, referredId_); 12966 } 12967 if (((bitField0_ & 0x00000002) == 0x00000002)) { 12968 size += com.google.protobuf.CodedOutputStream 12969 .computeBytesSize(2, name_); 12970 } 12971 if (((bitField0_ & 0x00000004) == 0x00000004)) { 12972 size += com.google.protobuf.CodedOutputStream 12973 .computeUInt32Size(3, dstSnapshotId_); 12974 } 12975 if (((bitField0_ & 0x00000008) == 0x00000008)) { 12976 size += com.google.protobuf.CodedOutputStream 12977 .computeUInt32Size(4, lastSnapshotId_); 12978 } 12979 size += getUnknownFields().getSerializedSize(); 12980 memoizedSerializedSize = size; 12981 return size; 12982 } 12983 12984 private static final long serialVersionUID = 0L; 12985 @java.lang.Override 12986 protected java.lang.Object writeReplace() 12987 throws java.io.ObjectStreamException { 12988 return super.writeReplace(); 12989 } 12990 12991 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 12992 com.google.protobuf.ByteString data) 12993 throws com.google.protobuf.InvalidProtocolBufferException { 12994 return PARSER.parseFrom(data); 12995 } 12996 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 12997 com.google.protobuf.ByteString data, 12998 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 12999 throws com.google.protobuf.InvalidProtocolBufferException { 13000 return PARSER.parseFrom(data, extensionRegistry); 13001 } 13002 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom(byte[] data) 13003 throws com.google.protobuf.InvalidProtocolBufferException { 13004 return PARSER.parseFrom(data); 13005 } 13006 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 13007 byte[] data, 13008 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13009 throws com.google.protobuf.InvalidProtocolBufferException { 13010 return PARSER.parseFrom(data, extensionRegistry); 13011 } 13012 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom(java.io.InputStream input) 13013 throws java.io.IOException { 13014 return PARSER.parseFrom(input); 13015 } 13016 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 13017 java.io.InputStream input, 13018 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13019 throws java.io.IOException { 13020 return PARSER.parseFrom(input, extensionRegistry); 13021 } 13022 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseDelimitedFrom(java.io.InputStream input) 13023 throws java.io.IOException { 13024 return PARSER.parseDelimitedFrom(input); 13025 } 13026 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseDelimitedFrom( 13027 java.io.InputStream input, 13028 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13029 throws java.io.IOException { 13030 return PARSER.parseDelimitedFrom(input, extensionRegistry); 13031 } 13032 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 13033 com.google.protobuf.CodedInputStream input) 13034 throws java.io.IOException { 13035 return PARSER.parseFrom(input); 13036 } 13037 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parseFrom( 13038 com.google.protobuf.CodedInputStream input, 13039 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13040 throws java.io.IOException { 13041 return PARSER.parseFrom(input, extensionRegistry); 13042 } 13043 13044 public static Builder newBuilder() { return Builder.create(); } 13045 public Builder newBuilderForType() { return newBuilder(); } 13046 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference prototype) { 13047 return newBuilder().mergeFrom(prototype); 13048 } 13049 public Builder toBuilder() { return newBuilder(this); } 13050 13051 @java.lang.Override 13052 protected Builder newBuilderForType( 13053 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13054 Builder builder = new Builder(parent); 13055 return builder; 13056 } 13057 /** 13058 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference} 13059 */ 13060 public static final class Builder extends 13061 com.google.protobuf.GeneratedMessage.Builder<Builder> 13062 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReferenceOrBuilder { 13063 public static final com.google.protobuf.Descriptors.Descriptor 13064 getDescriptor() { 13065 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 13066 } 13067 13068 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13069 internalGetFieldAccessorTable() { 13070 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable 13071 .ensureFieldAccessorsInitialized( 13072 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.Builder.class); 13073 } 13074 13075 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.newBuilder() 13076 private Builder() { 13077 maybeForceBuilderInitialization(); 13078 } 13079 13080 private Builder( 13081 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13082 super(parent); 13083 maybeForceBuilderInitialization(); 13084 } 13085 private void maybeForceBuilderInitialization() { 13086 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 13087 } 13088 } 13089 private static Builder create() { 13090 return new Builder(); 13091 } 13092 13093 public Builder clear() { 13094 super.clear(); 13095 referredId_ = 0L; 13096 bitField0_ = (bitField0_ & ~0x00000001); 13097 name_ = com.google.protobuf.ByteString.EMPTY; 13098 bitField0_ = (bitField0_ & ~0x00000002); 13099 dstSnapshotId_ = 0; 13100 bitField0_ = (bitField0_ & ~0x00000004); 13101 lastSnapshotId_ = 0; 13102 bitField0_ = (bitField0_ & ~0x00000008); 13103 return this; 13104 } 13105 13106 public Builder clone() { 13107 return create().mergeFrom(buildPartial()); 13108 } 13109 13110 public com.google.protobuf.Descriptors.Descriptor 13111 getDescriptorForType() { 13112 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 13113 } 13114 13115 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference getDefaultInstanceForType() { 13116 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.getDefaultInstance(); 13117 } 13118 13119 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference build() { 13120 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference result = buildPartial(); 13121 if (!result.isInitialized()) { 13122 throw newUninitializedMessageException(result); 13123 } 13124 return result; 13125 } 13126 13127 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference buildPartial() { 13128 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference(this); 13129 int from_bitField0_ = bitField0_; 13130 int to_bitField0_ = 0; 13131 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 13132 to_bitField0_ |= 0x00000001; 13133 } 13134 result.referredId_ = referredId_; 13135 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 13136 to_bitField0_ |= 0x00000002; 13137 } 13138 result.name_ = name_; 13139 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 13140 to_bitField0_ |= 0x00000004; 13141 } 13142 result.dstSnapshotId_ = dstSnapshotId_; 13143 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 13144 to_bitField0_ |= 0x00000008; 13145 } 13146 result.lastSnapshotId_ = lastSnapshotId_; 13147 result.bitField0_ = to_bitField0_; 13148 onBuilt(); 13149 return result; 13150 } 13151 13152 public Builder mergeFrom(com.google.protobuf.Message other) { 13153 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference) { 13154 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference)other); 13155 } else { 13156 super.mergeFrom(other); 13157 return this; 13158 } 13159 } 13160 13161 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference other) { 13162 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference.getDefaultInstance()) return this; 13163 if (other.hasReferredId()) { 13164 setReferredId(other.getReferredId()); 13165 } 13166 if (other.hasName()) { 13167 setName(other.getName()); 13168 } 13169 if (other.hasDstSnapshotId()) { 13170 setDstSnapshotId(other.getDstSnapshotId()); 13171 } 13172 if (other.hasLastSnapshotId()) { 13173 setLastSnapshotId(other.getLastSnapshotId()); 13174 } 13175 this.mergeUnknownFields(other.getUnknownFields()); 13176 return this; 13177 } 13178 13179 public final boolean isInitialized() { 13180 return true; 13181 } 13182 13183 public Builder mergeFrom( 13184 com.google.protobuf.CodedInputStream input, 13185 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13186 throws java.io.IOException { 13187 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference parsedMessage = null; 13188 try { 13189 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 13190 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13191 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.INodeReference) e.getUnfinishedMessage(); 13192 throw e; 13193 } finally { 13194 if (parsedMessage != null) { 13195 mergeFrom(parsedMessage); 13196 } 13197 } 13198 return this; 13199 } 13200 private int bitField0_; 13201 13202 // optional uint64 referredId = 1; 13203 private long referredId_ ; 13204 /** 13205 * <code>optional uint64 referredId = 1;</code> 13206 * 13207 * <pre> 13208 * id of the referred inode 13209 * </pre> 13210 */ 13211 public boolean hasReferredId() { 13212 return ((bitField0_ & 0x00000001) == 0x00000001); 13213 } 13214 /** 13215 * <code>optional uint64 referredId = 1;</code> 13216 * 13217 * <pre> 13218 * id of the referred inode 13219 * </pre> 13220 */ 13221 public long getReferredId() { 13222 return referredId_; 13223 } 13224 /** 13225 * <code>optional uint64 referredId = 1;</code> 13226 * 13227 * <pre> 13228 * id of the referred inode 13229 * </pre> 13230 */ 13231 public Builder setReferredId(long value) { 13232 bitField0_ |= 0x00000001; 13233 referredId_ = value; 13234 onChanged(); 13235 return this; 13236 } 13237 /** 13238 * <code>optional uint64 referredId = 1;</code> 13239 * 13240 * <pre> 13241 * id of the referred inode 13242 * </pre> 13243 */ 13244 public Builder clearReferredId() { 13245 bitField0_ = (bitField0_ & ~0x00000001); 13246 referredId_ = 0L; 13247 onChanged(); 13248 return this; 13249 } 13250 13251 // optional bytes name = 2; 13252 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 13253 /** 13254 * <code>optional bytes name = 2;</code> 13255 * 13256 * <pre> 13257 * local name recorded in WithName 13258 * </pre> 13259 */ 13260 public boolean hasName() { 13261 return ((bitField0_ & 0x00000002) == 0x00000002); 13262 } 13263 /** 13264 * <code>optional bytes name = 2;</code> 13265 * 13266 * <pre> 13267 * local name recorded in WithName 13268 * </pre> 13269 */ 13270 public com.google.protobuf.ByteString getName() { 13271 return name_; 13272 } 13273 /** 13274 * <code>optional bytes name = 2;</code> 13275 * 13276 * <pre> 13277 * local name recorded in WithName 13278 * </pre> 13279 */ 13280 public Builder setName(com.google.protobuf.ByteString value) { 13281 if (value == null) { 13282 throw new NullPointerException(); 13283 } 13284 bitField0_ |= 0x00000002; 13285 name_ = value; 13286 onChanged(); 13287 return this; 13288 } 13289 /** 13290 * <code>optional bytes name = 2;</code> 13291 * 13292 * <pre> 13293 * local name recorded in WithName 13294 * </pre> 13295 */ 13296 public Builder clearName() { 13297 bitField0_ = (bitField0_ & ~0x00000002); 13298 name_ = getDefaultInstance().getName(); 13299 onChanged(); 13300 return this; 13301 } 13302 13303 // optional uint32 dstSnapshotId = 3; 13304 private int dstSnapshotId_ ; 13305 /** 13306 * <code>optional uint32 dstSnapshotId = 3;</code> 13307 * 13308 * <pre> 13309 * recorded in DstReference 13310 * </pre> 13311 */ 13312 public boolean hasDstSnapshotId() { 13313 return ((bitField0_ & 0x00000004) == 0x00000004); 13314 } 13315 /** 13316 * <code>optional uint32 dstSnapshotId = 3;</code> 13317 * 13318 * <pre> 13319 * recorded in DstReference 13320 * </pre> 13321 */ 13322 public int getDstSnapshotId() { 13323 return dstSnapshotId_; 13324 } 13325 /** 13326 * <code>optional uint32 dstSnapshotId = 3;</code> 13327 * 13328 * <pre> 13329 * recorded in DstReference 13330 * </pre> 13331 */ 13332 public Builder setDstSnapshotId(int value) { 13333 bitField0_ |= 0x00000004; 13334 dstSnapshotId_ = value; 13335 onChanged(); 13336 return this; 13337 } 13338 /** 13339 * <code>optional uint32 dstSnapshotId = 3;</code> 13340 * 13341 * <pre> 13342 * recorded in DstReference 13343 * </pre> 13344 */ 13345 public Builder clearDstSnapshotId() { 13346 bitField0_ = (bitField0_ & ~0x00000004); 13347 dstSnapshotId_ = 0; 13348 onChanged(); 13349 return this; 13350 } 13351 13352 // optional uint32 lastSnapshotId = 4; 13353 private int lastSnapshotId_ ; 13354 /** 13355 * <code>optional uint32 lastSnapshotId = 4;</code> 13356 * 13357 * <pre> 13358 * recorded in WithName 13359 * </pre> 13360 */ 13361 public boolean hasLastSnapshotId() { 13362 return ((bitField0_ & 0x00000008) == 0x00000008); 13363 } 13364 /** 13365 * <code>optional uint32 lastSnapshotId = 4;</code> 13366 * 13367 * <pre> 13368 * recorded in WithName 13369 * </pre> 13370 */ 13371 public int getLastSnapshotId() { 13372 return lastSnapshotId_; 13373 } 13374 /** 13375 * <code>optional uint32 lastSnapshotId = 4;</code> 13376 * 13377 * <pre> 13378 * recorded in WithName 13379 * </pre> 13380 */ 13381 public Builder setLastSnapshotId(int value) { 13382 bitField0_ |= 0x00000008; 13383 lastSnapshotId_ = value; 13384 onChanged(); 13385 return this; 13386 } 13387 /** 13388 * <code>optional uint32 lastSnapshotId = 4;</code> 13389 * 13390 * <pre> 13391 * recorded in WithName 13392 * </pre> 13393 */ 13394 public Builder clearLastSnapshotId() { 13395 bitField0_ = (bitField0_ & ~0x00000008); 13396 lastSnapshotId_ = 0; 13397 onChanged(); 13398 return this; 13399 } 13400 13401 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference) 13402 } 13403 13404 static { 13405 defaultInstance = new INodeReference(true); 13406 defaultInstance.initFields(); 13407 } 13408 13409 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeReferenceSection.INodeReference) 13410 } 13411 13412 private void initFields() { 13413 } 13414 private byte memoizedIsInitialized = -1; 13415 public final boolean isInitialized() { 13416 byte isInitialized = memoizedIsInitialized; 13417 if (isInitialized != -1) return isInitialized == 1; 13418 13419 memoizedIsInitialized = 1; 13420 return true; 13421 } 13422 13423 public void writeTo(com.google.protobuf.CodedOutputStream output) 13424 throws java.io.IOException { 13425 getSerializedSize(); 13426 getUnknownFields().writeTo(output); 13427 } 13428 13429 private int memoizedSerializedSize = -1; 13430 public int getSerializedSize() { 13431 int size = memoizedSerializedSize; 13432 if (size != -1) return size; 13433 13434 size = 0; 13435 size += getUnknownFields().getSerializedSize(); 13436 memoizedSerializedSize = size; 13437 return size; 13438 } 13439 13440 private static final long serialVersionUID = 0L; 13441 @java.lang.Override 13442 protected java.lang.Object writeReplace() 13443 throws java.io.ObjectStreamException { 13444 return super.writeReplace(); 13445 } 13446 13447 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13448 com.google.protobuf.ByteString data) 13449 throws com.google.protobuf.InvalidProtocolBufferException { 13450 return PARSER.parseFrom(data); 13451 } 13452 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13453 com.google.protobuf.ByteString data, 13454 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13455 throws com.google.protobuf.InvalidProtocolBufferException { 13456 return PARSER.parseFrom(data, extensionRegistry); 13457 } 13458 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom(byte[] data) 13459 throws com.google.protobuf.InvalidProtocolBufferException { 13460 return PARSER.parseFrom(data); 13461 } 13462 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13463 byte[] data, 13464 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13465 throws com.google.protobuf.InvalidProtocolBufferException { 13466 return PARSER.parseFrom(data, extensionRegistry); 13467 } 13468 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom(java.io.InputStream input) 13469 throws java.io.IOException { 13470 return PARSER.parseFrom(input); 13471 } 13472 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13473 java.io.InputStream input, 13474 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13475 throws java.io.IOException { 13476 return PARSER.parseFrom(input, extensionRegistry); 13477 } 13478 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseDelimitedFrom(java.io.InputStream input) 13479 throws java.io.IOException { 13480 return PARSER.parseDelimitedFrom(input); 13481 } 13482 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseDelimitedFrom( 13483 java.io.InputStream input, 13484 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13485 throws java.io.IOException { 13486 return PARSER.parseDelimitedFrom(input, extensionRegistry); 13487 } 13488 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13489 com.google.protobuf.CodedInputStream input) 13490 throws java.io.IOException { 13491 return PARSER.parseFrom(input); 13492 } 13493 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parseFrom( 13494 com.google.protobuf.CodedInputStream input, 13495 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13496 throws java.io.IOException { 13497 return PARSER.parseFrom(input, extensionRegistry); 13498 } 13499 13500 public static Builder newBuilder() { return Builder.create(); } 13501 public Builder newBuilderForType() { return newBuilder(); } 13502 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection prototype) { 13503 return newBuilder().mergeFrom(prototype); 13504 } 13505 public Builder toBuilder() { return newBuilder(this); } 13506 13507 @java.lang.Override 13508 protected Builder newBuilderForType( 13509 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13510 Builder builder = new Builder(parent); 13511 return builder; 13512 } 13513 /** 13514 * Protobuf type {@code hadoop.hdfs.fsimage.INodeReferenceSection} 13515 */ 13516 public static final class Builder extends 13517 com.google.protobuf.GeneratedMessage.Builder<Builder> 13518 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSectionOrBuilder { 13519 public static final com.google.protobuf.Descriptors.Descriptor 13520 getDescriptor() { 13521 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 13522 } 13523 13524 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13525 internalGetFieldAccessorTable() { 13526 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable 13527 .ensureFieldAccessorsInitialized( 13528 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.Builder.class); 13529 } 13530 13531 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.newBuilder() 13532 private Builder() { 13533 maybeForceBuilderInitialization(); 13534 } 13535 13536 private Builder( 13537 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 13538 super(parent); 13539 maybeForceBuilderInitialization(); 13540 } 13541 private void maybeForceBuilderInitialization() { 13542 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 13543 } 13544 } 13545 private static Builder create() { 13546 return new Builder(); 13547 } 13548 13549 public Builder clear() { 13550 super.clear(); 13551 return this; 13552 } 13553 13554 public Builder clone() { 13555 return create().mergeFrom(buildPartial()); 13556 } 13557 13558 public com.google.protobuf.Descriptors.Descriptor 13559 getDescriptorForType() { 13560 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 13561 } 13562 13563 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection getDefaultInstanceForType() { 13564 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.getDefaultInstance(); 13565 } 13566 13567 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection build() { 13568 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection result = buildPartial(); 13569 if (!result.isInitialized()) { 13570 throw newUninitializedMessageException(result); 13571 } 13572 return result; 13573 } 13574 13575 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection buildPartial() { 13576 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection(this); 13577 onBuilt(); 13578 return result; 13579 } 13580 13581 public Builder mergeFrom(com.google.protobuf.Message other) { 13582 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection) { 13583 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection)other); 13584 } else { 13585 super.mergeFrom(other); 13586 return this; 13587 } 13588 } 13589 13590 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection other) { 13591 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection.getDefaultInstance()) return this; 13592 this.mergeUnknownFields(other.getUnknownFields()); 13593 return this; 13594 } 13595 13596 public final boolean isInitialized() { 13597 return true; 13598 } 13599 13600 public Builder mergeFrom( 13601 com.google.protobuf.CodedInputStream input, 13602 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13603 throws java.io.IOException { 13604 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection parsedMessage = null; 13605 try { 13606 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 13607 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13608 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeReferenceSection) e.getUnfinishedMessage(); 13609 throw e; 13610 } finally { 13611 if (parsedMessage != null) { 13612 mergeFrom(parsedMessage); 13613 } 13614 } 13615 return this; 13616 } 13617 13618 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.INodeReferenceSection) 13619 } 13620 13621 static { 13622 defaultInstance = new INodeReferenceSection(true); 13623 defaultInstance.initFields(); 13624 } 13625 13626 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.INodeReferenceSection) 13627 } 13628 13629 public interface SnapshotSectionOrBuilder 13630 extends com.google.protobuf.MessageOrBuilder { 13631 13632 // optional uint32 snapshotCounter = 1; 13633 /** 13634 * <code>optional uint32 snapshotCounter = 1;</code> 13635 */ 13636 boolean hasSnapshotCounter(); 13637 /** 13638 * <code>optional uint32 snapshotCounter = 1;</code> 13639 */ 13640 int getSnapshotCounter(); 13641 13642 // repeated uint64 snapshottableDir = 2 [packed = true]; 13643 /** 13644 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13645 */ 13646 java.util.List<java.lang.Long> getSnapshottableDirList(); 13647 /** 13648 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13649 */ 13650 int getSnapshottableDirCount(); 13651 /** 13652 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 13653 */ 13654 long getSnapshottableDir(int index); 13655 13656 // optional uint32 numSnapshots = 3; 13657 /** 13658 * <code>optional uint32 numSnapshots = 3;</code> 13659 * 13660 * <pre> 13661 * total number of snapshots 13662 * </pre> 13663 */ 13664 boolean hasNumSnapshots(); 13665 /** 13666 * <code>optional uint32 numSnapshots = 3;</code> 13667 * 13668 * <pre> 13669 * total number of snapshots 13670 * </pre> 13671 */ 13672 int getNumSnapshots(); 13673 } 13674 /** 13675 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection} 13676 * 13677 * <pre> 13678 ** 13679 * This section records the information about snapshot 13680 * NAME: SNAPSHOT 13681 * </pre> 13682 */ 13683 public static final class SnapshotSection extends 13684 com.google.protobuf.GeneratedMessage 13685 implements SnapshotSectionOrBuilder { 13686 // Use SnapshotSection.newBuilder() to construct. 13687 private SnapshotSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13688 super(builder); 13689 this.unknownFields = builder.getUnknownFields(); 13690 } 13691 private SnapshotSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13692 13693 private static final SnapshotSection defaultInstance; 13694 public static SnapshotSection getDefaultInstance() { 13695 return defaultInstance; 13696 } 13697 13698 public SnapshotSection getDefaultInstanceForType() { 13699 return defaultInstance; 13700 } 13701 13702 private final com.google.protobuf.UnknownFieldSet unknownFields; 13703 @java.lang.Override 13704 public final com.google.protobuf.UnknownFieldSet 13705 getUnknownFields() { 13706 return this.unknownFields; 13707 } 13708 private SnapshotSection( 13709 com.google.protobuf.CodedInputStream input, 13710 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13711 throws com.google.protobuf.InvalidProtocolBufferException { 13712 initFields(); 13713 int mutable_bitField0_ = 0; 13714 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13715 com.google.protobuf.UnknownFieldSet.newBuilder(); 13716 try { 13717 boolean done = false; 13718 while (!done) { 13719 int tag = input.readTag(); 13720 switch (tag) { 13721 case 0: 13722 done = true; 13723 break; 13724 default: { 13725 if (!parseUnknownField(input, unknownFields, 13726 extensionRegistry, tag)) { 13727 done = true; 13728 } 13729 break; 13730 } 13731 case 8: { 13732 bitField0_ |= 0x00000001; 13733 snapshotCounter_ = input.readUInt32(); 13734 break; 13735 } 13736 case 16: { 13737 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 13738 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(); 13739 mutable_bitField0_ |= 0x00000002; 13740 } 13741 snapshottableDir_.add(input.readUInt64()); 13742 break; 13743 } 13744 case 18: { 13745 int length = input.readRawVarint32(); 13746 int limit = input.pushLimit(length); 13747 if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) { 13748 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(); 13749 mutable_bitField0_ |= 0x00000002; 13750 } 13751 while (input.getBytesUntilLimit() > 0) { 13752 snapshottableDir_.add(input.readUInt64()); 13753 } 13754 input.popLimit(limit); 13755 break; 13756 } 13757 case 24: { 13758 bitField0_ |= 0x00000002; 13759 numSnapshots_ = input.readUInt32(); 13760 break; 13761 } 13762 } 13763 } 13764 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13765 throw e.setUnfinishedMessage(this); 13766 } catch (java.io.IOException e) { 13767 throw new com.google.protobuf.InvalidProtocolBufferException( 13768 e.getMessage()).setUnfinishedMessage(this); 13769 } finally { 13770 if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { 13771 snapshottableDir_ = java.util.Collections.unmodifiableList(snapshottableDir_); 13772 } 13773 this.unknownFields = unknownFields.build(); 13774 makeExtensionsImmutable(); 13775 } 13776 } 13777 public static final com.google.protobuf.Descriptors.Descriptor 13778 getDescriptor() { 13779 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 13780 } 13781 13782 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13783 internalGetFieldAccessorTable() { 13784 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable 13785 .ensureFieldAccessorsInitialized( 13786 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Builder.class); 13787 } 13788 13789 public static com.google.protobuf.Parser<SnapshotSection> PARSER = 13790 new com.google.protobuf.AbstractParser<SnapshotSection>() { 13791 public SnapshotSection parsePartialFrom( 13792 com.google.protobuf.CodedInputStream input, 13793 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13794 throws com.google.protobuf.InvalidProtocolBufferException { 13795 return new SnapshotSection(input, extensionRegistry); 13796 } 13797 }; 13798 13799 @java.lang.Override 13800 public com.google.protobuf.Parser<SnapshotSection> getParserForType() { 13801 return PARSER; 13802 } 13803 13804 public interface SnapshotOrBuilder 13805 extends com.google.protobuf.MessageOrBuilder { 13806 13807 // optional uint32 snapshotId = 1; 13808 /** 13809 * <code>optional uint32 snapshotId = 1;</code> 13810 */ 13811 boolean hasSnapshotId(); 13812 /** 13813 * <code>optional uint32 snapshotId = 1;</code> 13814 */ 13815 int getSnapshotId(); 13816 13817 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 13818 /** 13819 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13820 * 13821 * <pre> 13822 * Snapshot root 13823 * </pre> 13824 */ 13825 boolean hasRoot(); 13826 /** 13827 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13828 * 13829 * <pre> 13830 * Snapshot root 13831 * </pre> 13832 */ 13833 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot(); 13834 /** 13835 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13836 * 13837 * <pre> 13838 * Snapshot root 13839 * </pre> 13840 */ 13841 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder(); 13842 } 13843 /** 13844 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection.Snapshot} 13845 */ 13846 public static final class Snapshot extends 13847 com.google.protobuf.GeneratedMessage 13848 implements SnapshotOrBuilder { 13849 // Use Snapshot.newBuilder() to construct. 13850 private Snapshot(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 13851 super(builder); 13852 this.unknownFields = builder.getUnknownFields(); 13853 } 13854 private Snapshot(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 13855 13856 private static final Snapshot defaultInstance; 13857 public static Snapshot getDefaultInstance() { 13858 return defaultInstance; 13859 } 13860 13861 public Snapshot getDefaultInstanceForType() { 13862 return defaultInstance; 13863 } 13864 13865 private final com.google.protobuf.UnknownFieldSet unknownFields; 13866 @java.lang.Override 13867 public final com.google.protobuf.UnknownFieldSet 13868 getUnknownFields() { 13869 return this.unknownFields; 13870 } 13871 private Snapshot( 13872 com.google.protobuf.CodedInputStream input, 13873 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13874 throws com.google.protobuf.InvalidProtocolBufferException { 13875 initFields(); 13876 int mutable_bitField0_ = 0; 13877 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 13878 com.google.protobuf.UnknownFieldSet.newBuilder(); 13879 try { 13880 boolean done = false; 13881 while (!done) { 13882 int tag = input.readTag(); 13883 switch (tag) { 13884 case 0: 13885 done = true; 13886 break; 13887 default: { 13888 if (!parseUnknownField(input, unknownFields, 13889 extensionRegistry, tag)) { 13890 done = true; 13891 } 13892 break; 13893 } 13894 case 8: { 13895 bitField0_ |= 0x00000001; 13896 snapshotId_ = input.readUInt32(); 13897 break; 13898 } 13899 case 18: { 13900 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder subBuilder = null; 13901 if (((bitField0_ & 0x00000002) == 0x00000002)) { 13902 subBuilder = root_.toBuilder(); 13903 } 13904 root_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.PARSER, extensionRegistry); 13905 if (subBuilder != null) { 13906 subBuilder.mergeFrom(root_); 13907 root_ = subBuilder.buildPartial(); 13908 } 13909 bitField0_ |= 0x00000002; 13910 break; 13911 } 13912 } 13913 } 13914 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 13915 throw e.setUnfinishedMessage(this); 13916 } catch (java.io.IOException e) { 13917 throw new com.google.protobuf.InvalidProtocolBufferException( 13918 e.getMessage()).setUnfinishedMessage(this); 13919 } finally { 13920 this.unknownFields = unknownFields.build(); 13921 makeExtensionsImmutable(); 13922 } 13923 } 13924 public static final com.google.protobuf.Descriptors.Descriptor 13925 getDescriptor() { 13926 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 13927 } 13928 13929 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 13930 internalGetFieldAccessorTable() { 13931 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable 13932 .ensureFieldAccessorsInitialized( 13933 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.Builder.class); 13934 } 13935 13936 public static com.google.protobuf.Parser<Snapshot> PARSER = 13937 new com.google.protobuf.AbstractParser<Snapshot>() { 13938 public Snapshot parsePartialFrom( 13939 com.google.protobuf.CodedInputStream input, 13940 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 13941 throws com.google.protobuf.InvalidProtocolBufferException { 13942 return new Snapshot(input, extensionRegistry); 13943 } 13944 }; 13945 13946 @java.lang.Override 13947 public com.google.protobuf.Parser<Snapshot> getParserForType() { 13948 return PARSER; 13949 } 13950 13951 private int bitField0_; 13952 // optional uint32 snapshotId = 1; 13953 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 13954 private int snapshotId_; 13955 /** 13956 * <code>optional uint32 snapshotId = 1;</code> 13957 */ 13958 public boolean hasSnapshotId() { 13959 return ((bitField0_ & 0x00000001) == 0x00000001); 13960 } 13961 /** 13962 * <code>optional uint32 snapshotId = 1;</code> 13963 */ 13964 public int getSnapshotId() { 13965 return snapshotId_; 13966 } 13967 13968 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 13969 public static final int ROOT_FIELD_NUMBER = 2; 13970 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode root_; 13971 /** 13972 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13973 * 13974 * <pre> 13975 * Snapshot root 13976 * </pre> 13977 */ 13978 public boolean hasRoot() { 13979 return ((bitField0_ & 0x00000002) == 0x00000002); 13980 } 13981 /** 13982 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13983 * 13984 * <pre> 13985 * Snapshot root 13986 * </pre> 13987 */ 13988 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot() { 13989 return root_; 13990 } 13991 /** 13992 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 13993 * 13994 * <pre> 13995 * Snapshot root 13996 * </pre> 13997 */ 13998 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder() { 13999 return root_; 14000 } 14001 14002 private void initFields() { 14003 snapshotId_ = 0; 14004 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 14005 } 14006 private byte memoizedIsInitialized = -1; 14007 public final boolean isInitialized() { 14008 byte isInitialized = memoizedIsInitialized; 14009 if (isInitialized != -1) return isInitialized == 1; 14010 14011 if (hasRoot()) { 14012 if (!getRoot().isInitialized()) { 14013 memoizedIsInitialized = 0; 14014 return false; 14015 } 14016 } 14017 memoizedIsInitialized = 1; 14018 return true; 14019 } 14020 14021 public void writeTo(com.google.protobuf.CodedOutputStream output) 14022 throws java.io.IOException { 14023 getSerializedSize(); 14024 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14025 output.writeUInt32(1, snapshotId_); 14026 } 14027 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14028 output.writeMessage(2, root_); 14029 } 14030 getUnknownFields().writeTo(output); 14031 } 14032 14033 private int memoizedSerializedSize = -1; 14034 public int getSerializedSize() { 14035 int size = memoizedSerializedSize; 14036 if (size != -1) return size; 14037 14038 size = 0; 14039 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14040 size += com.google.protobuf.CodedOutputStream 14041 .computeUInt32Size(1, snapshotId_); 14042 } 14043 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14044 size += com.google.protobuf.CodedOutputStream 14045 .computeMessageSize(2, root_); 14046 } 14047 size += getUnknownFields().getSerializedSize(); 14048 memoizedSerializedSize = size; 14049 return size; 14050 } 14051 14052 private static final long serialVersionUID = 0L; 14053 @java.lang.Override 14054 protected java.lang.Object writeReplace() 14055 throws java.io.ObjectStreamException { 14056 return super.writeReplace(); 14057 } 14058 14059 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14060 com.google.protobuf.ByteString data) 14061 throws com.google.protobuf.InvalidProtocolBufferException { 14062 return PARSER.parseFrom(data); 14063 } 14064 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14065 com.google.protobuf.ByteString data, 14066 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14067 throws com.google.protobuf.InvalidProtocolBufferException { 14068 return PARSER.parseFrom(data, extensionRegistry); 14069 } 14070 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom(byte[] data) 14071 throws com.google.protobuf.InvalidProtocolBufferException { 14072 return PARSER.parseFrom(data); 14073 } 14074 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14075 byte[] data, 14076 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14077 throws com.google.protobuf.InvalidProtocolBufferException { 14078 return PARSER.parseFrom(data, extensionRegistry); 14079 } 14080 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom(java.io.InputStream input) 14081 throws java.io.IOException { 14082 return PARSER.parseFrom(input); 14083 } 14084 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14085 java.io.InputStream input, 14086 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14087 throws java.io.IOException { 14088 return PARSER.parseFrom(input, extensionRegistry); 14089 } 14090 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseDelimitedFrom(java.io.InputStream input) 14091 throws java.io.IOException { 14092 return PARSER.parseDelimitedFrom(input); 14093 } 14094 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseDelimitedFrom( 14095 java.io.InputStream input, 14096 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14097 throws java.io.IOException { 14098 return PARSER.parseDelimitedFrom(input, extensionRegistry); 14099 } 14100 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14101 com.google.protobuf.CodedInputStream input) 14102 throws java.io.IOException { 14103 return PARSER.parseFrom(input); 14104 } 14105 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parseFrom( 14106 com.google.protobuf.CodedInputStream input, 14107 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14108 throws java.io.IOException { 14109 return PARSER.parseFrom(input, extensionRegistry); 14110 } 14111 14112 public static Builder newBuilder() { return Builder.create(); } 14113 public Builder newBuilderForType() { return newBuilder(); } 14114 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot prototype) { 14115 return newBuilder().mergeFrom(prototype); 14116 } 14117 public Builder toBuilder() { return newBuilder(this); } 14118 14119 @java.lang.Override 14120 protected Builder newBuilderForType( 14121 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14122 Builder builder = new Builder(parent); 14123 return builder; 14124 } 14125 /** 14126 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection.Snapshot} 14127 */ 14128 public static final class Builder extends 14129 com.google.protobuf.GeneratedMessage.Builder<Builder> 14130 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.SnapshotOrBuilder { 14131 public static final com.google.protobuf.Descriptors.Descriptor 14132 getDescriptor() { 14133 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 14134 } 14135 14136 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14137 internalGetFieldAccessorTable() { 14138 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable 14139 .ensureFieldAccessorsInitialized( 14140 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.Builder.class); 14141 } 14142 14143 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.newBuilder() 14144 private Builder() { 14145 maybeForceBuilderInitialization(); 14146 } 14147 14148 private Builder( 14149 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14150 super(parent); 14151 maybeForceBuilderInitialization(); 14152 } 14153 private void maybeForceBuilderInitialization() { 14154 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 14155 getRootFieldBuilder(); 14156 } 14157 } 14158 private static Builder create() { 14159 return new Builder(); 14160 } 14161 14162 public Builder clear() { 14163 super.clear(); 14164 snapshotId_ = 0; 14165 bitField0_ = (bitField0_ & ~0x00000001); 14166 if (rootBuilder_ == null) { 14167 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 14168 } else { 14169 rootBuilder_.clear(); 14170 } 14171 bitField0_ = (bitField0_ & ~0x00000002); 14172 return this; 14173 } 14174 14175 public Builder clone() { 14176 return create().mergeFrom(buildPartial()); 14177 } 14178 14179 public com.google.protobuf.Descriptors.Descriptor 14180 getDescriptorForType() { 14181 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 14182 } 14183 14184 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot getDefaultInstanceForType() { 14185 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.getDefaultInstance(); 14186 } 14187 14188 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot build() { 14189 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot result = buildPartial(); 14190 if (!result.isInitialized()) { 14191 throw newUninitializedMessageException(result); 14192 } 14193 return result; 14194 } 14195 14196 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot buildPartial() { 14197 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot(this); 14198 int from_bitField0_ = bitField0_; 14199 int to_bitField0_ = 0; 14200 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 14201 to_bitField0_ |= 0x00000001; 14202 } 14203 result.snapshotId_ = snapshotId_; 14204 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 14205 to_bitField0_ |= 0x00000002; 14206 } 14207 if (rootBuilder_ == null) { 14208 result.root_ = root_; 14209 } else { 14210 result.root_ = rootBuilder_.build(); 14211 } 14212 result.bitField0_ = to_bitField0_; 14213 onBuilt(); 14214 return result; 14215 } 14216 14217 public Builder mergeFrom(com.google.protobuf.Message other) { 14218 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot) { 14219 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot)other); 14220 } else { 14221 super.mergeFrom(other); 14222 return this; 14223 } 14224 } 14225 14226 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot other) { 14227 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot.getDefaultInstance()) return this; 14228 if (other.hasSnapshotId()) { 14229 setSnapshotId(other.getSnapshotId()); 14230 } 14231 if (other.hasRoot()) { 14232 mergeRoot(other.getRoot()); 14233 } 14234 this.mergeUnknownFields(other.getUnknownFields()); 14235 return this; 14236 } 14237 14238 public final boolean isInitialized() { 14239 if (hasRoot()) { 14240 if (!getRoot().isInitialized()) { 14241 14242 return false; 14243 } 14244 } 14245 return true; 14246 } 14247 14248 public Builder mergeFrom( 14249 com.google.protobuf.CodedInputStream input, 14250 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14251 throws java.io.IOException { 14252 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot parsedMessage = null; 14253 try { 14254 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 14255 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 14256 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Snapshot) e.getUnfinishedMessage(); 14257 throw e; 14258 } finally { 14259 if (parsedMessage != null) { 14260 mergeFrom(parsedMessage); 14261 } 14262 } 14263 return this; 14264 } 14265 private int bitField0_; 14266 14267 // optional uint32 snapshotId = 1; 14268 private int snapshotId_ ; 14269 /** 14270 * <code>optional uint32 snapshotId = 1;</code> 14271 */ 14272 public boolean hasSnapshotId() { 14273 return ((bitField0_ & 0x00000001) == 0x00000001); 14274 } 14275 /** 14276 * <code>optional uint32 snapshotId = 1;</code> 14277 */ 14278 public int getSnapshotId() { 14279 return snapshotId_; 14280 } 14281 /** 14282 * <code>optional uint32 snapshotId = 1;</code> 14283 */ 14284 public Builder setSnapshotId(int value) { 14285 bitField0_ |= 0x00000001; 14286 snapshotId_ = value; 14287 onChanged(); 14288 return this; 14289 } 14290 /** 14291 * <code>optional uint32 snapshotId = 1;</code> 14292 */ 14293 public Builder clearSnapshotId() { 14294 bitField0_ = (bitField0_ & ~0x00000001); 14295 snapshotId_ = 0; 14296 onChanged(); 14297 return this; 14298 } 14299 14300 // optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2; 14301 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 14302 private com.google.protobuf.SingleFieldBuilder< 14303 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder> rootBuilder_; 14304 /** 14305 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14306 * 14307 * <pre> 14308 * Snapshot root 14309 * </pre> 14310 */ 14311 public boolean hasRoot() { 14312 return ((bitField0_ & 0x00000002) == 0x00000002); 14313 } 14314 /** 14315 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14316 * 14317 * <pre> 14318 * Snapshot root 14319 * </pre> 14320 */ 14321 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode getRoot() { 14322 if (rootBuilder_ == null) { 14323 return root_; 14324 } else { 14325 return rootBuilder_.getMessage(); 14326 } 14327 } 14328 /** 14329 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14330 * 14331 * <pre> 14332 * Snapshot root 14333 * </pre> 14334 */ 14335 public Builder setRoot(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode value) { 14336 if (rootBuilder_ == null) { 14337 if (value == null) { 14338 throw new NullPointerException(); 14339 } 14340 root_ = value; 14341 onChanged(); 14342 } else { 14343 rootBuilder_.setMessage(value); 14344 } 14345 bitField0_ |= 0x00000002; 14346 return this; 14347 } 14348 /** 14349 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14350 * 14351 * <pre> 14352 * Snapshot root 14353 * </pre> 14354 */ 14355 public Builder setRoot( 14356 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder builderForValue) { 14357 if (rootBuilder_ == null) { 14358 root_ = builderForValue.build(); 14359 onChanged(); 14360 } else { 14361 rootBuilder_.setMessage(builderForValue.build()); 14362 } 14363 bitField0_ |= 0x00000002; 14364 return this; 14365 } 14366 /** 14367 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14368 * 14369 * <pre> 14370 * Snapshot root 14371 * </pre> 14372 */ 14373 public Builder mergeRoot(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode value) { 14374 if (rootBuilder_ == null) { 14375 if (((bitField0_ & 0x00000002) == 0x00000002) && 14376 root_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance()) { 14377 root_ = 14378 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.newBuilder(root_).mergeFrom(value).buildPartial(); 14379 } else { 14380 root_ = value; 14381 } 14382 onChanged(); 14383 } else { 14384 rootBuilder_.mergeFrom(value); 14385 } 14386 bitField0_ |= 0x00000002; 14387 return this; 14388 } 14389 /** 14390 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14391 * 14392 * <pre> 14393 * Snapshot root 14394 * </pre> 14395 */ 14396 public Builder clearRoot() { 14397 if (rootBuilder_ == null) { 14398 root_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.getDefaultInstance(); 14399 onChanged(); 14400 } else { 14401 rootBuilder_.clear(); 14402 } 14403 bitField0_ = (bitField0_ & ~0x00000002); 14404 return this; 14405 } 14406 /** 14407 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14408 * 14409 * <pre> 14410 * Snapshot root 14411 * </pre> 14412 */ 14413 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder getRootBuilder() { 14414 bitField0_ |= 0x00000002; 14415 onChanged(); 14416 return getRootFieldBuilder().getBuilder(); 14417 } 14418 /** 14419 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14420 * 14421 * <pre> 14422 * Snapshot root 14423 * </pre> 14424 */ 14425 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder getRootOrBuilder() { 14426 if (rootBuilder_ != null) { 14427 return rootBuilder_.getMessageOrBuilder(); 14428 } else { 14429 return root_; 14430 } 14431 } 14432 /** 14433 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INode root = 2;</code> 14434 * 14435 * <pre> 14436 * Snapshot root 14437 * </pre> 14438 */ 14439 private com.google.protobuf.SingleFieldBuilder< 14440 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder> 14441 getRootFieldBuilder() { 14442 if (rootBuilder_ == null) { 14443 rootBuilder_ = new com.google.protobuf.SingleFieldBuilder< 14444 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INode.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeOrBuilder>( 14445 root_, 14446 getParentForChildren(), 14447 isClean()); 14448 root_ = null; 14449 } 14450 return rootBuilder_; 14451 } 14452 14453 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotSection.Snapshot) 14454 } 14455 14456 static { 14457 defaultInstance = new Snapshot(true); 14458 defaultInstance.initFields(); 14459 } 14460 14461 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotSection.Snapshot) 14462 } 14463 14464 private int bitField0_; 14465 // optional uint32 snapshotCounter = 1; 14466 public static final int SNAPSHOTCOUNTER_FIELD_NUMBER = 1; 14467 private int snapshotCounter_; 14468 /** 14469 * <code>optional uint32 snapshotCounter = 1;</code> 14470 */ 14471 public boolean hasSnapshotCounter() { 14472 return ((bitField0_ & 0x00000001) == 0x00000001); 14473 } 14474 /** 14475 * <code>optional uint32 snapshotCounter = 1;</code> 14476 */ 14477 public int getSnapshotCounter() { 14478 return snapshotCounter_; 14479 } 14480 14481 // repeated uint64 snapshottableDir = 2 [packed = true]; 14482 public static final int SNAPSHOTTABLEDIR_FIELD_NUMBER = 2; 14483 private java.util.List<java.lang.Long> snapshottableDir_; 14484 /** 14485 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14486 */ 14487 public java.util.List<java.lang.Long> 14488 getSnapshottableDirList() { 14489 return snapshottableDir_; 14490 } 14491 /** 14492 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14493 */ 14494 public int getSnapshottableDirCount() { 14495 return snapshottableDir_.size(); 14496 } 14497 /** 14498 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14499 */ 14500 public long getSnapshottableDir(int index) { 14501 return snapshottableDir_.get(index); 14502 } 14503 private int snapshottableDirMemoizedSerializedSize = -1; 14504 14505 // optional uint32 numSnapshots = 3; 14506 public static final int NUMSNAPSHOTS_FIELD_NUMBER = 3; 14507 private int numSnapshots_; 14508 /** 14509 * <code>optional uint32 numSnapshots = 3;</code> 14510 * 14511 * <pre> 14512 * total number of snapshots 14513 * </pre> 14514 */ 14515 public boolean hasNumSnapshots() { 14516 return ((bitField0_ & 0x00000002) == 0x00000002); 14517 } 14518 /** 14519 * <code>optional uint32 numSnapshots = 3;</code> 14520 * 14521 * <pre> 14522 * total number of snapshots 14523 * </pre> 14524 */ 14525 public int getNumSnapshots() { 14526 return numSnapshots_; 14527 } 14528 14529 private void initFields() { 14530 snapshotCounter_ = 0; 14531 snapshottableDir_ = java.util.Collections.emptyList(); 14532 numSnapshots_ = 0; 14533 } 14534 private byte memoizedIsInitialized = -1; 14535 public final boolean isInitialized() { 14536 byte isInitialized = memoizedIsInitialized; 14537 if (isInitialized != -1) return isInitialized == 1; 14538 14539 memoizedIsInitialized = 1; 14540 return true; 14541 } 14542 14543 public void writeTo(com.google.protobuf.CodedOutputStream output) 14544 throws java.io.IOException { 14545 getSerializedSize(); 14546 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14547 output.writeUInt32(1, snapshotCounter_); 14548 } 14549 if (getSnapshottableDirList().size() > 0) { 14550 output.writeRawVarint32(18); 14551 output.writeRawVarint32(snapshottableDirMemoizedSerializedSize); 14552 } 14553 for (int i = 0; i < snapshottableDir_.size(); i++) { 14554 output.writeUInt64NoTag(snapshottableDir_.get(i)); 14555 } 14556 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14557 output.writeUInt32(3, numSnapshots_); 14558 } 14559 getUnknownFields().writeTo(output); 14560 } 14561 14562 private int memoizedSerializedSize = -1; 14563 public int getSerializedSize() { 14564 int size = memoizedSerializedSize; 14565 if (size != -1) return size; 14566 14567 size = 0; 14568 if (((bitField0_ & 0x00000001) == 0x00000001)) { 14569 size += com.google.protobuf.CodedOutputStream 14570 .computeUInt32Size(1, snapshotCounter_); 14571 } 14572 { 14573 int dataSize = 0; 14574 for (int i = 0; i < snapshottableDir_.size(); i++) { 14575 dataSize += com.google.protobuf.CodedOutputStream 14576 .computeUInt64SizeNoTag(snapshottableDir_.get(i)); 14577 } 14578 size += dataSize; 14579 if (!getSnapshottableDirList().isEmpty()) { 14580 size += 1; 14581 size += com.google.protobuf.CodedOutputStream 14582 .computeInt32SizeNoTag(dataSize); 14583 } 14584 snapshottableDirMemoizedSerializedSize = dataSize; 14585 } 14586 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14587 size += com.google.protobuf.CodedOutputStream 14588 .computeUInt32Size(3, numSnapshots_); 14589 } 14590 size += getUnknownFields().getSerializedSize(); 14591 memoizedSerializedSize = size; 14592 return size; 14593 } 14594 14595 private static final long serialVersionUID = 0L; 14596 @java.lang.Override 14597 protected java.lang.Object writeReplace() 14598 throws java.io.ObjectStreamException { 14599 return super.writeReplace(); 14600 } 14601 14602 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14603 com.google.protobuf.ByteString data) 14604 throws com.google.protobuf.InvalidProtocolBufferException { 14605 return PARSER.parseFrom(data); 14606 } 14607 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14608 com.google.protobuf.ByteString data, 14609 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14610 throws com.google.protobuf.InvalidProtocolBufferException { 14611 return PARSER.parseFrom(data, extensionRegistry); 14612 } 14613 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom(byte[] data) 14614 throws com.google.protobuf.InvalidProtocolBufferException { 14615 return PARSER.parseFrom(data); 14616 } 14617 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14618 byte[] data, 14619 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14620 throws com.google.protobuf.InvalidProtocolBufferException { 14621 return PARSER.parseFrom(data, extensionRegistry); 14622 } 14623 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom(java.io.InputStream input) 14624 throws java.io.IOException { 14625 return PARSER.parseFrom(input); 14626 } 14627 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14628 java.io.InputStream input, 14629 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14630 throws java.io.IOException { 14631 return PARSER.parseFrom(input, extensionRegistry); 14632 } 14633 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseDelimitedFrom(java.io.InputStream input) 14634 throws java.io.IOException { 14635 return PARSER.parseDelimitedFrom(input); 14636 } 14637 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseDelimitedFrom( 14638 java.io.InputStream input, 14639 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14640 throws java.io.IOException { 14641 return PARSER.parseDelimitedFrom(input, extensionRegistry); 14642 } 14643 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14644 com.google.protobuf.CodedInputStream input) 14645 throws java.io.IOException { 14646 return PARSER.parseFrom(input); 14647 } 14648 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parseFrom( 14649 com.google.protobuf.CodedInputStream input, 14650 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14651 throws java.io.IOException { 14652 return PARSER.parseFrom(input, extensionRegistry); 14653 } 14654 14655 public static Builder newBuilder() { return Builder.create(); } 14656 public Builder newBuilderForType() { return newBuilder(); } 14657 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection prototype) { 14658 return newBuilder().mergeFrom(prototype); 14659 } 14660 public Builder toBuilder() { return newBuilder(this); } 14661 14662 @java.lang.Override 14663 protected Builder newBuilderForType( 14664 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14665 Builder builder = new Builder(parent); 14666 return builder; 14667 } 14668 /** 14669 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotSection} 14670 * 14671 * <pre> 14672 ** 14673 * This section records the information about snapshot 14674 * NAME: SNAPSHOT 14675 * </pre> 14676 */ 14677 public static final class Builder extends 14678 com.google.protobuf.GeneratedMessage.Builder<Builder> 14679 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSectionOrBuilder { 14680 public static final com.google.protobuf.Descriptors.Descriptor 14681 getDescriptor() { 14682 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 14683 } 14684 14685 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 14686 internalGetFieldAccessorTable() { 14687 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable 14688 .ensureFieldAccessorsInitialized( 14689 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.Builder.class); 14690 } 14691 14692 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.newBuilder() 14693 private Builder() { 14694 maybeForceBuilderInitialization(); 14695 } 14696 14697 private Builder( 14698 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 14699 super(parent); 14700 maybeForceBuilderInitialization(); 14701 } 14702 private void maybeForceBuilderInitialization() { 14703 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 14704 } 14705 } 14706 private static Builder create() { 14707 return new Builder(); 14708 } 14709 14710 public Builder clear() { 14711 super.clear(); 14712 snapshotCounter_ = 0; 14713 bitField0_ = (bitField0_ & ~0x00000001); 14714 snapshottableDir_ = java.util.Collections.emptyList(); 14715 bitField0_ = (bitField0_ & ~0x00000002); 14716 numSnapshots_ = 0; 14717 bitField0_ = (bitField0_ & ~0x00000004); 14718 return this; 14719 } 14720 14721 public Builder clone() { 14722 return create().mergeFrom(buildPartial()); 14723 } 14724 14725 public com.google.protobuf.Descriptors.Descriptor 14726 getDescriptorForType() { 14727 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 14728 } 14729 14730 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection getDefaultInstanceForType() { 14731 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.getDefaultInstance(); 14732 } 14733 14734 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection build() { 14735 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection result = buildPartial(); 14736 if (!result.isInitialized()) { 14737 throw newUninitializedMessageException(result); 14738 } 14739 return result; 14740 } 14741 14742 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection buildPartial() { 14743 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection(this); 14744 int from_bitField0_ = bitField0_; 14745 int to_bitField0_ = 0; 14746 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 14747 to_bitField0_ |= 0x00000001; 14748 } 14749 result.snapshotCounter_ = snapshotCounter_; 14750 if (((bitField0_ & 0x00000002) == 0x00000002)) { 14751 snapshottableDir_ = java.util.Collections.unmodifiableList(snapshottableDir_); 14752 bitField0_ = (bitField0_ & ~0x00000002); 14753 } 14754 result.snapshottableDir_ = snapshottableDir_; 14755 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 14756 to_bitField0_ |= 0x00000002; 14757 } 14758 result.numSnapshots_ = numSnapshots_; 14759 result.bitField0_ = to_bitField0_; 14760 onBuilt(); 14761 return result; 14762 } 14763 14764 public Builder mergeFrom(com.google.protobuf.Message other) { 14765 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection) { 14766 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection)other); 14767 } else { 14768 super.mergeFrom(other); 14769 return this; 14770 } 14771 } 14772 14773 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection other) { 14774 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection.getDefaultInstance()) return this; 14775 if (other.hasSnapshotCounter()) { 14776 setSnapshotCounter(other.getSnapshotCounter()); 14777 } 14778 if (!other.snapshottableDir_.isEmpty()) { 14779 if (snapshottableDir_.isEmpty()) { 14780 snapshottableDir_ = other.snapshottableDir_; 14781 bitField0_ = (bitField0_ & ~0x00000002); 14782 } else { 14783 ensureSnapshottableDirIsMutable(); 14784 snapshottableDir_.addAll(other.snapshottableDir_); 14785 } 14786 onChanged(); 14787 } 14788 if (other.hasNumSnapshots()) { 14789 setNumSnapshots(other.getNumSnapshots()); 14790 } 14791 this.mergeUnknownFields(other.getUnknownFields()); 14792 return this; 14793 } 14794 14795 public final boolean isInitialized() { 14796 return true; 14797 } 14798 14799 public Builder mergeFrom( 14800 com.google.protobuf.CodedInputStream input, 14801 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 14802 throws java.io.IOException { 14803 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection parsedMessage = null; 14804 try { 14805 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 14806 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 14807 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotSection) e.getUnfinishedMessage(); 14808 throw e; 14809 } finally { 14810 if (parsedMessage != null) { 14811 mergeFrom(parsedMessage); 14812 } 14813 } 14814 return this; 14815 } 14816 private int bitField0_; 14817 14818 // optional uint32 snapshotCounter = 1; 14819 private int snapshotCounter_ ; 14820 /** 14821 * <code>optional uint32 snapshotCounter = 1;</code> 14822 */ 14823 public boolean hasSnapshotCounter() { 14824 return ((bitField0_ & 0x00000001) == 0x00000001); 14825 } 14826 /** 14827 * <code>optional uint32 snapshotCounter = 1;</code> 14828 */ 14829 public int getSnapshotCounter() { 14830 return snapshotCounter_; 14831 } 14832 /** 14833 * <code>optional uint32 snapshotCounter = 1;</code> 14834 */ 14835 public Builder setSnapshotCounter(int value) { 14836 bitField0_ |= 0x00000001; 14837 snapshotCounter_ = value; 14838 onChanged(); 14839 return this; 14840 } 14841 /** 14842 * <code>optional uint32 snapshotCounter = 1;</code> 14843 */ 14844 public Builder clearSnapshotCounter() { 14845 bitField0_ = (bitField0_ & ~0x00000001); 14846 snapshotCounter_ = 0; 14847 onChanged(); 14848 return this; 14849 } 14850 14851 // repeated uint64 snapshottableDir = 2 [packed = true]; 14852 private java.util.List<java.lang.Long> snapshottableDir_ = java.util.Collections.emptyList(); 14853 private void ensureSnapshottableDirIsMutable() { 14854 if (!((bitField0_ & 0x00000002) == 0x00000002)) { 14855 snapshottableDir_ = new java.util.ArrayList<java.lang.Long>(snapshottableDir_); 14856 bitField0_ |= 0x00000002; 14857 } 14858 } 14859 /** 14860 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14861 */ 14862 public java.util.List<java.lang.Long> 14863 getSnapshottableDirList() { 14864 return java.util.Collections.unmodifiableList(snapshottableDir_); 14865 } 14866 /** 14867 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14868 */ 14869 public int getSnapshottableDirCount() { 14870 return snapshottableDir_.size(); 14871 } 14872 /** 14873 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14874 */ 14875 public long getSnapshottableDir(int index) { 14876 return snapshottableDir_.get(index); 14877 } 14878 /** 14879 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14880 */ 14881 public Builder setSnapshottableDir( 14882 int index, long value) { 14883 ensureSnapshottableDirIsMutable(); 14884 snapshottableDir_.set(index, value); 14885 onChanged(); 14886 return this; 14887 } 14888 /** 14889 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14890 */ 14891 public Builder addSnapshottableDir(long value) { 14892 ensureSnapshottableDirIsMutable(); 14893 snapshottableDir_.add(value); 14894 onChanged(); 14895 return this; 14896 } 14897 /** 14898 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14899 */ 14900 public Builder addAllSnapshottableDir( 14901 java.lang.Iterable<? extends java.lang.Long> values) { 14902 ensureSnapshottableDirIsMutable(); 14903 super.addAll(values, snapshottableDir_); 14904 onChanged(); 14905 return this; 14906 } 14907 /** 14908 * <code>repeated uint64 snapshottableDir = 2 [packed = true];</code> 14909 */ 14910 public Builder clearSnapshottableDir() { 14911 snapshottableDir_ = java.util.Collections.emptyList(); 14912 bitField0_ = (bitField0_ & ~0x00000002); 14913 onChanged(); 14914 return this; 14915 } 14916 14917 // optional uint32 numSnapshots = 3; 14918 private int numSnapshots_ ; 14919 /** 14920 * <code>optional uint32 numSnapshots = 3;</code> 14921 * 14922 * <pre> 14923 * total number of snapshots 14924 * </pre> 14925 */ 14926 public boolean hasNumSnapshots() { 14927 return ((bitField0_ & 0x00000004) == 0x00000004); 14928 } 14929 /** 14930 * <code>optional uint32 numSnapshots = 3;</code> 14931 * 14932 * <pre> 14933 * total number of snapshots 14934 * </pre> 14935 */ 14936 public int getNumSnapshots() { 14937 return numSnapshots_; 14938 } 14939 /** 14940 * <code>optional uint32 numSnapshots = 3;</code> 14941 * 14942 * <pre> 14943 * total number of snapshots 14944 * </pre> 14945 */ 14946 public Builder setNumSnapshots(int value) { 14947 bitField0_ |= 0x00000004; 14948 numSnapshots_ = value; 14949 onChanged(); 14950 return this; 14951 } 14952 /** 14953 * <code>optional uint32 numSnapshots = 3;</code> 14954 * 14955 * <pre> 14956 * total number of snapshots 14957 * </pre> 14958 */ 14959 public Builder clearNumSnapshots() { 14960 bitField0_ = (bitField0_ & ~0x00000004); 14961 numSnapshots_ = 0; 14962 onChanged(); 14963 return this; 14964 } 14965 14966 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotSection) 14967 } 14968 14969 static { 14970 defaultInstance = new SnapshotSection(true); 14971 defaultInstance.initFields(); 14972 } 14973 14974 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotSection) 14975 } 14976 14977 public interface SnapshotDiffSectionOrBuilder 14978 extends com.google.protobuf.MessageOrBuilder { 14979 } 14980 /** 14981 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection} 14982 * 14983 * <pre> 14984 ** 14985 * This section records information about snapshot diffs 14986 * NAME: SNAPSHOT_DIFF 14987 * </pre> 14988 */ 14989 public static final class SnapshotDiffSection extends 14990 com.google.protobuf.GeneratedMessage 14991 implements SnapshotDiffSectionOrBuilder { 14992 // Use SnapshotDiffSection.newBuilder() to construct. 14993 private SnapshotDiffSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 14994 super(builder); 14995 this.unknownFields = builder.getUnknownFields(); 14996 } 14997 private SnapshotDiffSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 14998 14999 private static final SnapshotDiffSection defaultInstance; 15000 public static SnapshotDiffSection getDefaultInstance() { 15001 return defaultInstance; 15002 } 15003 15004 public SnapshotDiffSection getDefaultInstanceForType() { 15005 return defaultInstance; 15006 } 15007 15008 private final com.google.protobuf.UnknownFieldSet unknownFields; 15009 @java.lang.Override 15010 public final com.google.protobuf.UnknownFieldSet 15011 getUnknownFields() { 15012 return this.unknownFields; 15013 } 15014 private SnapshotDiffSection( 15015 com.google.protobuf.CodedInputStream input, 15016 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15017 throws com.google.protobuf.InvalidProtocolBufferException { 15018 initFields(); 15019 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15020 com.google.protobuf.UnknownFieldSet.newBuilder(); 15021 try { 15022 boolean done = false; 15023 while (!done) { 15024 int tag = input.readTag(); 15025 switch (tag) { 15026 case 0: 15027 done = true; 15028 break; 15029 default: { 15030 if (!parseUnknownField(input, unknownFields, 15031 extensionRegistry, tag)) { 15032 done = true; 15033 } 15034 break; 15035 } 15036 } 15037 } 15038 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15039 throw e.setUnfinishedMessage(this); 15040 } catch (java.io.IOException e) { 15041 throw new com.google.protobuf.InvalidProtocolBufferException( 15042 e.getMessage()).setUnfinishedMessage(this); 15043 } finally { 15044 this.unknownFields = unknownFields.build(); 15045 makeExtensionsImmutable(); 15046 } 15047 } 15048 public static final com.google.protobuf.Descriptors.Descriptor 15049 getDescriptor() { 15050 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 15051 } 15052 15053 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15054 internalGetFieldAccessorTable() { 15055 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable 15056 .ensureFieldAccessorsInitialized( 15057 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.Builder.class); 15058 } 15059 15060 public static com.google.protobuf.Parser<SnapshotDiffSection> PARSER = 15061 new com.google.protobuf.AbstractParser<SnapshotDiffSection>() { 15062 public SnapshotDiffSection parsePartialFrom( 15063 com.google.protobuf.CodedInputStream input, 15064 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15065 throws com.google.protobuf.InvalidProtocolBufferException { 15066 return new SnapshotDiffSection(input, extensionRegistry); 15067 } 15068 }; 15069 15070 @java.lang.Override 15071 public com.google.protobuf.Parser<SnapshotDiffSection> getParserForType() { 15072 return PARSER; 15073 } 15074 15075 public interface CreatedListEntryOrBuilder 15076 extends com.google.protobuf.MessageOrBuilder { 15077 15078 // optional bytes name = 1; 15079 /** 15080 * <code>optional bytes name = 1;</code> 15081 */ 15082 boolean hasName(); 15083 /** 15084 * <code>optional bytes name = 1;</code> 15085 */ 15086 com.google.protobuf.ByteString getName(); 15087 } 15088 /** 15089 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry} 15090 */ 15091 public static final class CreatedListEntry extends 15092 com.google.protobuf.GeneratedMessage 15093 implements CreatedListEntryOrBuilder { 15094 // Use CreatedListEntry.newBuilder() to construct. 15095 private CreatedListEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 15096 super(builder); 15097 this.unknownFields = builder.getUnknownFields(); 15098 } 15099 private CreatedListEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 15100 15101 private static final CreatedListEntry defaultInstance; 15102 public static CreatedListEntry getDefaultInstance() { 15103 return defaultInstance; 15104 } 15105 15106 public CreatedListEntry getDefaultInstanceForType() { 15107 return defaultInstance; 15108 } 15109 15110 private final com.google.protobuf.UnknownFieldSet unknownFields; 15111 @java.lang.Override 15112 public final com.google.protobuf.UnknownFieldSet 15113 getUnknownFields() { 15114 return this.unknownFields; 15115 } 15116 private CreatedListEntry( 15117 com.google.protobuf.CodedInputStream input, 15118 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15119 throws com.google.protobuf.InvalidProtocolBufferException { 15120 initFields(); 15121 int mutable_bitField0_ = 0; 15122 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15123 com.google.protobuf.UnknownFieldSet.newBuilder(); 15124 try { 15125 boolean done = false; 15126 while (!done) { 15127 int tag = input.readTag(); 15128 switch (tag) { 15129 case 0: 15130 done = true; 15131 break; 15132 default: { 15133 if (!parseUnknownField(input, unknownFields, 15134 extensionRegistry, tag)) { 15135 done = true; 15136 } 15137 break; 15138 } 15139 case 10: { 15140 bitField0_ |= 0x00000001; 15141 name_ = input.readBytes(); 15142 break; 15143 } 15144 } 15145 } 15146 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15147 throw e.setUnfinishedMessage(this); 15148 } catch (java.io.IOException e) { 15149 throw new com.google.protobuf.InvalidProtocolBufferException( 15150 e.getMessage()).setUnfinishedMessage(this); 15151 } finally { 15152 this.unknownFields = unknownFields.build(); 15153 makeExtensionsImmutable(); 15154 } 15155 } 15156 public static final com.google.protobuf.Descriptors.Descriptor 15157 getDescriptor() { 15158 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 15159 } 15160 15161 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15162 internalGetFieldAccessorTable() { 15163 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable 15164 .ensureFieldAccessorsInitialized( 15165 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.Builder.class); 15166 } 15167 15168 public static com.google.protobuf.Parser<CreatedListEntry> PARSER = 15169 new com.google.protobuf.AbstractParser<CreatedListEntry>() { 15170 public CreatedListEntry parsePartialFrom( 15171 com.google.protobuf.CodedInputStream input, 15172 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15173 throws com.google.protobuf.InvalidProtocolBufferException { 15174 return new CreatedListEntry(input, extensionRegistry); 15175 } 15176 }; 15177 15178 @java.lang.Override 15179 public com.google.protobuf.Parser<CreatedListEntry> getParserForType() { 15180 return PARSER; 15181 } 15182 15183 private int bitField0_; 15184 // optional bytes name = 1; 15185 public static final int NAME_FIELD_NUMBER = 1; 15186 private com.google.protobuf.ByteString name_; 15187 /** 15188 * <code>optional bytes name = 1;</code> 15189 */ 15190 public boolean hasName() { 15191 return ((bitField0_ & 0x00000001) == 0x00000001); 15192 } 15193 /** 15194 * <code>optional bytes name = 1;</code> 15195 */ 15196 public com.google.protobuf.ByteString getName() { 15197 return name_; 15198 } 15199 15200 private void initFields() { 15201 name_ = com.google.protobuf.ByteString.EMPTY; 15202 } 15203 private byte memoizedIsInitialized = -1; 15204 public final boolean isInitialized() { 15205 byte isInitialized = memoizedIsInitialized; 15206 if (isInitialized != -1) return isInitialized == 1; 15207 15208 memoizedIsInitialized = 1; 15209 return true; 15210 } 15211 15212 public void writeTo(com.google.protobuf.CodedOutputStream output) 15213 throws java.io.IOException { 15214 getSerializedSize(); 15215 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15216 output.writeBytes(1, name_); 15217 } 15218 getUnknownFields().writeTo(output); 15219 } 15220 15221 private int memoizedSerializedSize = -1; 15222 public int getSerializedSize() { 15223 int size = memoizedSerializedSize; 15224 if (size != -1) return size; 15225 15226 size = 0; 15227 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15228 size += com.google.protobuf.CodedOutputStream 15229 .computeBytesSize(1, name_); 15230 } 15231 size += getUnknownFields().getSerializedSize(); 15232 memoizedSerializedSize = size; 15233 return size; 15234 } 15235 15236 private static final long serialVersionUID = 0L; 15237 @java.lang.Override 15238 protected java.lang.Object writeReplace() 15239 throws java.io.ObjectStreamException { 15240 return super.writeReplace(); 15241 } 15242 15243 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15244 com.google.protobuf.ByteString data) 15245 throws com.google.protobuf.InvalidProtocolBufferException { 15246 return PARSER.parseFrom(data); 15247 } 15248 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15249 com.google.protobuf.ByteString data, 15250 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15251 throws com.google.protobuf.InvalidProtocolBufferException { 15252 return PARSER.parseFrom(data, extensionRegistry); 15253 } 15254 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom(byte[] data) 15255 throws com.google.protobuf.InvalidProtocolBufferException { 15256 return PARSER.parseFrom(data); 15257 } 15258 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15259 byte[] data, 15260 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15261 throws com.google.protobuf.InvalidProtocolBufferException { 15262 return PARSER.parseFrom(data, extensionRegistry); 15263 } 15264 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom(java.io.InputStream input) 15265 throws java.io.IOException { 15266 return PARSER.parseFrom(input); 15267 } 15268 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15269 java.io.InputStream input, 15270 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15271 throws java.io.IOException { 15272 return PARSER.parseFrom(input, extensionRegistry); 15273 } 15274 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseDelimitedFrom(java.io.InputStream input) 15275 throws java.io.IOException { 15276 return PARSER.parseDelimitedFrom(input); 15277 } 15278 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseDelimitedFrom( 15279 java.io.InputStream input, 15280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15281 throws java.io.IOException { 15282 return PARSER.parseDelimitedFrom(input, extensionRegistry); 15283 } 15284 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15285 com.google.protobuf.CodedInputStream input) 15286 throws java.io.IOException { 15287 return PARSER.parseFrom(input); 15288 } 15289 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parseFrom( 15290 com.google.protobuf.CodedInputStream input, 15291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15292 throws java.io.IOException { 15293 return PARSER.parseFrom(input, extensionRegistry); 15294 } 15295 15296 public static Builder newBuilder() { return Builder.create(); } 15297 public Builder newBuilderForType() { return newBuilder(); } 15298 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry prototype) { 15299 return newBuilder().mergeFrom(prototype); 15300 } 15301 public Builder toBuilder() { return newBuilder(this); } 15302 15303 @java.lang.Override 15304 protected Builder newBuilderForType( 15305 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15306 Builder builder = new Builder(parent); 15307 return builder; 15308 } 15309 /** 15310 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry} 15311 */ 15312 public static final class Builder extends 15313 com.google.protobuf.GeneratedMessage.Builder<Builder> 15314 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntryOrBuilder { 15315 public static final com.google.protobuf.Descriptors.Descriptor 15316 getDescriptor() { 15317 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 15318 } 15319 15320 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15321 internalGetFieldAccessorTable() { 15322 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable 15323 .ensureFieldAccessorsInitialized( 15324 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.Builder.class); 15325 } 15326 15327 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.newBuilder() 15328 private Builder() { 15329 maybeForceBuilderInitialization(); 15330 } 15331 15332 private Builder( 15333 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 15334 super(parent); 15335 maybeForceBuilderInitialization(); 15336 } 15337 private void maybeForceBuilderInitialization() { 15338 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 15339 } 15340 } 15341 private static Builder create() { 15342 return new Builder(); 15343 } 15344 15345 public Builder clear() { 15346 super.clear(); 15347 name_ = com.google.protobuf.ByteString.EMPTY; 15348 bitField0_ = (bitField0_ & ~0x00000001); 15349 return this; 15350 } 15351 15352 public Builder clone() { 15353 return create().mergeFrom(buildPartial()); 15354 } 15355 15356 public com.google.protobuf.Descriptors.Descriptor 15357 getDescriptorForType() { 15358 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 15359 } 15360 15361 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry getDefaultInstanceForType() { 15362 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.getDefaultInstance(); 15363 } 15364 15365 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry build() { 15366 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry result = buildPartial(); 15367 if (!result.isInitialized()) { 15368 throw newUninitializedMessageException(result); 15369 } 15370 return result; 15371 } 15372 15373 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry buildPartial() { 15374 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry(this); 15375 int from_bitField0_ = bitField0_; 15376 int to_bitField0_ = 0; 15377 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 15378 to_bitField0_ |= 0x00000001; 15379 } 15380 result.name_ = name_; 15381 result.bitField0_ = to_bitField0_; 15382 onBuilt(); 15383 return result; 15384 } 15385 15386 public Builder mergeFrom(com.google.protobuf.Message other) { 15387 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry) { 15388 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry)other); 15389 } else { 15390 super.mergeFrom(other); 15391 return this; 15392 } 15393 } 15394 15395 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry other) { 15396 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry.getDefaultInstance()) return this; 15397 if (other.hasName()) { 15398 setName(other.getName()); 15399 } 15400 this.mergeUnknownFields(other.getUnknownFields()); 15401 return this; 15402 } 15403 15404 public final boolean isInitialized() { 15405 return true; 15406 } 15407 15408 public Builder mergeFrom( 15409 com.google.protobuf.CodedInputStream input, 15410 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15411 throws java.io.IOException { 15412 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry parsedMessage = null; 15413 try { 15414 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 15415 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15416 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.CreatedListEntry) e.getUnfinishedMessage(); 15417 throw e; 15418 } finally { 15419 if (parsedMessage != null) { 15420 mergeFrom(parsedMessage); 15421 } 15422 } 15423 return this; 15424 } 15425 private int bitField0_; 15426 15427 // optional bytes name = 1; 15428 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 15429 /** 15430 * <code>optional bytes name = 1;</code> 15431 */ 15432 public boolean hasName() { 15433 return ((bitField0_ & 0x00000001) == 0x00000001); 15434 } 15435 /** 15436 * <code>optional bytes name = 1;</code> 15437 */ 15438 public com.google.protobuf.ByteString getName() { 15439 return name_; 15440 } 15441 /** 15442 * <code>optional bytes name = 1;</code> 15443 */ 15444 public Builder setName(com.google.protobuf.ByteString value) { 15445 if (value == null) { 15446 throw new NullPointerException(); 15447 } 15448 bitField0_ |= 0x00000001; 15449 name_ = value; 15450 onChanged(); 15451 return this; 15452 } 15453 /** 15454 * <code>optional bytes name = 1;</code> 15455 */ 15456 public Builder clearName() { 15457 bitField0_ = (bitField0_ & ~0x00000001); 15458 name_ = getDefaultInstance().getName(); 15459 onChanged(); 15460 return this; 15461 } 15462 15463 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry) 15464 } 15465 15466 static { 15467 defaultInstance = new CreatedListEntry(true); 15468 defaultInstance.initFields(); 15469 } 15470 15471 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.CreatedListEntry) 15472 } 15473 15474 public interface DirectoryDiffOrBuilder 15475 extends com.google.protobuf.MessageOrBuilder { 15476 15477 // optional uint32 snapshotId = 1; 15478 /** 15479 * <code>optional uint32 snapshotId = 1;</code> 15480 */ 15481 boolean hasSnapshotId(); 15482 /** 15483 * <code>optional uint32 snapshotId = 1;</code> 15484 */ 15485 int getSnapshotId(); 15486 15487 // optional uint32 childrenSize = 2; 15488 /** 15489 * <code>optional uint32 childrenSize = 2;</code> 15490 */ 15491 boolean hasChildrenSize(); 15492 /** 15493 * <code>optional uint32 childrenSize = 2;</code> 15494 */ 15495 int getChildrenSize(); 15496 15497 // optional bool isSnapshotRoot = 3; 15498 /** 15499 * <code>optional bool isSnapshotRoot = 3;</code> 15500 */ 15501 boolean hasIsSnapshotRoot(); 15502 /** 15503 * <code>optional bool isSnapshotRoot = 3;</code> 15504 */ 15505 boolean getIsSnapshotRoot(); 15506 15507 // optional bytes name = 4; 15508 /** 15509 * <code>optional bytes name = 4;</code> 15510 */ 15511 boolean hasName(); 15512 /** 15513 * <code>optional bytes name = 4;</code> 15514 */ 15515 com.google.protobuf.ByteString getName(); 15516 15517 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 15518 /** 15519 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15520 */ 15521 boolean hasSnapshotCopy(); 15522 /** 15523 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15524 */ 15525 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy(); 15526 /** 15527 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15528 */ 15529 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder(); 15530 15531 // optional uint32 createdListSize = 6; 15532 /** 15533 * <code>optional uint32 createdListSize = 6;</code> 15534 */ 15535 boolean hasCreatedListSize(); 15536 /** 15537 * <code>optional uint32 createdListSize = 6;</code> 15538 */ 15539 int getCreatedListSize(); 15540 15541 // repeated uint64 deletedINode = 7 [packed = true]; 15542 /** 15543 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15544 * 15545 * <pre> 15546 * id of deleted inodes 15547 * </pre> 15548 */ 15549 java.util.List<java.lang.Long> getDeletedINodeList(); 15550 /** 15551 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15552 * 15553 * <pre> 15554 * id of deleted inodes 15555 * </pre> 15556 */ 15557 int getDeletedINodeCount(); 15558 /** 15559 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15560 * 15561 * <pre> 15562 * id of deleted inodes 15563 * </pre> 15564 */ 15565 long getDeletedINode(int index); 15566 15567 // repeated uint32 deletedINodeRef = 8 [packed = true]; 15568 /** 15569 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15570 * 15571 * <pre> 15572 * id of reference nodes in the deleted list 15573 * </pre> 15574 */ 15575 java.util.List<java.lang.Integer> getDeletedINodeRefList(); 15576 /** 15577 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15578 * 15579 * <pre> 15580 * id of reference nodes in the deleted list 15581 * </pre> 15582 */ 15583 int getDeletedINodeRefCount(); 15584 /** 15585 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15586 * 15587 * <pre> 15588 * id of reference nodes in the deleted list 15589 * </pre> 15590 */ 15591 int getDeletedINodeRef(int index); 15592 } 15593 /** 15594 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff} 15595 */ 15596 public static final class DirectoryDiff extends 15597 com.google.protobuf.GeneratedMessage 15598 implements DirectoryDiffOrBuilder { 15599 // Use DirectoryDiff.newBuilder() to construct. 15600 private DirectoryDiff(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 15601 super(builder); 15602 this.unknownFields = builder.getUnknownFields(); 15603 } 15604 private DirectoryDiff(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 15605 15606 private static final DirectoryDiff defaultInstance; 15607 public static DirectoryDiff getDefaultInstance() { 15608 return defaultInstance; 15609 } 15610 15611 public DirectoryDiff getDefaultInstanceForType() { 15612 return defaultInstance; 15613 } 15614 15615 private final com.google.protobuf.UnknownFieldSet unknownFields; 15616 @java.lang.Override 15617 public final com.google.protobuf.UnknownFieldSet 15618 getUnknownFields() { 15619 return this.unknownFields; 15620 } 15621 private DirectoryDiff( 15622 com.google.protobuf.CodedInputStream input, 15623 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15624 throws com.google.protobuf.InvalidProtocolBufferException { 15625 initFields(); 15626 int mutable_bitField0_ = 0; 15627 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 15628 com.google.protobuf.UnknownFieldSet.newBuilder(); 15629 try { 15630 boolean done = false; 15631 while (!done) { 15632 int tag = input.readTag(); 15633 switch (tag) { 15634 case 0: 15635 done = true; 15636 break; 15637 default: { 15638 if (!parseUnknownField(input, unknownFields, 15639 extensionRegistry, tag)) { 15640 done = true; 15641 } 15642 break; 15643 } 15644 case 8: { 15645 bitField0_ |= 0x00000001; 15646 snapshotId_ = input.readUInt32(); 15647 break; 15648 } 15649 case 16: { 15650 bitField0_ |= 0x00000002; 15651 childrenSize_ = input.readUInt32(); 15652 break; 15653 } 15654 case 24: { 15655 bitField0_ |= 0x00000004; 15656 isSnapshotRoot_ = input.readBool(); 15657 break; 15658 } 15659 case 34: { 15660 bitField0_ |= 0x00000008; 15661 name_ = input.readBytes(); 15662 break; 15663 } 15664 case 42: { 15665 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder subBuilder = null; 15666 if (((bitField0_ & 0x00000010) == 0x00000010)) { 15667 subBuilder = snapshotCopy_.toBuilder(); 15668 } 15669 snapshotCopy_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.PARSER, extensionRegistry); 15670 if (subBuilder != null) { 15671 subBuilder.mergeFrom(snapshotCopy_); 15672 snapshotCopy_ = subBuilder.buildPartial(); 15673 } 15674 bitField0_ |= 0x00000010; 15675 break; 15676 } 15677 case 48: { 15678 bitField0_ |= 0x00000020; 15679 createdListSize_ = input.readUInt32(); 15680 break; 15681 } 15682 case 56: { 15683 if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) { 15684 deletedINode_ = new java.util.ArrayList<java.lang.Long>(); 15685 mutable_bitField0_ |= 0x00000040; 15686 } 15687 deletedINode_.add(input.readUInt64()); 15688 break; 15689 } 15690 case 58: { 15691 int length = input.readRawVarint32(); 15692 int limit = input.pushLimit(length); 15693 if (!((mutable_bitField0_ & 0x00000040) == 0x00000040) && input.getBytesUntilLimit() > 0) { 15694 deletedINode_ = new java.util.ArrayList<java.lang.Long>(); 15695 mutable_bitField0_ |= 0x00000040; 15696 } 15697 while (input.getBytesUntilLimit() > 0) { 15698 deletedINode_.add(input.readUInt64()); 15699 } 15700 input.popLimit(limit); 15701 break; 15702 } 15703 case 64: { 15704 if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) { 15705 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(); 15706 mutable_bitField0_ |= 0x00000080; 15707 } 15708 deletedINodeRef_.add(input.readUInt32()); 15709 break; 15710 } 15711 case 66: { 15712 int length = input.readRawVarint32(); 15713 int limit = input.pushLimit(length); 15714 if (!((mutable_bitField0_ & 0x00000080) == 0x00000080) && input.getBytesUntilLimit() > 0) { 15715 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(); 15716 mutable_bitField0_ |= 0x00000080; 15717 } 15718 while (input.getBytesUntilLimit() > 0) { 15719 deletedINodeRef_.add(input.readUInt32()); 15720 } 15721 input.popLimit(limit); 15722 break; 15723 } 15724 } 15725 } 15726 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 15727 throw e.setUnfinishedMessage(this); 15728 } catch (java.io.IOException e) { 15729 throw new com.google.protobuf.InvalidProtocolBufferException( 15730 e.getMessage()).setUnfinishedMessage(this); 15731 } finally { 15732 if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) { 15733 deletedINode_ = java.util.Collections.unmodifiableList(deletedINode_); 15734 } 15735 if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) { 15736 deletedINodeRef_ = java.util.Collections.unmodifiableList(deletedINodeRef_); 15737 } 15738 this.unknownFields = unknownFields.build(); 15739 makeExtensionsImmutable(); 15740 } 15741 } 15742 public static final com.google.protobuf.Descriptors.Descriptor 15743 getDescriptor() { 15744 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 15745 } 15746 15747 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 15748 internalGetFieldAccessorTable() { 15749 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable 15750 .ensureFieldAccessorsInitialized( 15751 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.Builder.class); 15752 } 15753 15754 public static com.google.protobuf.Parser<DirectoryDiff> PARSER = 15755 new com.google.protobuf.AbstractParser<DirectoryDiff>() { 15756 public DirectoryDiff parsePartialFrom( 15757 com.google.protobuf.CodedInputStream input, 15758 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 15759 throws com.google.protobuf.InvalidProtocolBufferException { 15760 return new DirectoryDiff(input, extensionRegistry); 15761 } 15762 }; 15763 15764 @java.lang.Override 15765 public com.google.protobuf.Parser<DirectoryDiff> getParserForType() { 15766 return PARSER; 15767 } 15768 15769 private int bitField0_; 15770 // optional uint32 snapshotId = 1; 15771 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 15772 private int snapshotId_; 15773 /** 15774 * <code>optional uint32 snapshotId = 1;</code> 15775 */ 15776 public boolean hasSnapshotId() { 15777 return ((bitField0_ & 0x00000001) == 0x00000001); 15778 } 15779 /** 15780 * <code>optional uint32 snapshotId = 1;</code> 15781 */ 15782 public int getSnapshotId() { 15783 return snapshotId_; 15784 } 15785 15786 // optional uint32 childrenSize = 2; 15787 public static final int CHILDRENSIZE_FIELD_NUMBER = 2; 15788 private int childrenSize_; 15789 /** 15790 * <code>optional uint32 childrenSize = 2;</code> 15791 */ 15792 public boolean hasChildrenSize() { 15793 return ((bitField0_ & 0x00000002) == 0x00000002); 15794 } 15795 /** 15796 * <code>optional uint32 childrenSize = 2;</code> 15797 */ 15798 public int getChildrenSize() { 15799 return childrenSize_; 15800 } 15801 15802 // optional bool isSnapshotRoot = 3; 15803 public static final int ISSNAPSHOTROOT_FIELD_NUMBER = 3; 15804 private boolean isSnapshotRoot_; 15805 /** 15806 * <code>optional bool isSnapshotRoot = 3;</code> 15807 */ 15808 public boolean hasIsSnapshotRoot() { 15809 return ((bitField0_ & 0x00000004) == 0x00000004); 15810 } 15811 /** 15812 * <code>optional bool isSnapshotRoot = 3;</code> 15813 */ 15814 public boolean getIsSnapshotRoot() { 15815 return isSnapshotRoot_; 15816 } 15817 15818 // optional bytes name = 4; 15819 public static final int NAME_FIELD_NUMBER = 4; 15820 private com.google.protobuf.ByteString name_; 15821 /** 15822 * <code>optional bytes name = 4;</code> 15823 */ 15824 public boolean hasName() { 15825 return ((bitField0_ & 0x00000008) == 0x00000008); 15826 } 15827 /** 15828 * <code>optional bytes name = 4;</code> 15829 */ 15830 public com.google.protobuf.ByteString getName() { 15831 return name_; 15832 } 15833 15834 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 15835 public static final int SNAPSHOTCOPY_FIELD_NUMBER = 5; 15836 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory snapshotCopy_; 15837 /** 15838 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15839 */ 15840 public boolean hasSnapshotCopy() { 15841 return ((bitField0_ & 0x00000010) == 0x00000010); 15842 } 15843 /** 15844 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15845 */ 15846 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy() { 15847 return snapshotCopy_; 15848 } 15849 /** 15850 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 15851 */ 15852 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder() { 15853 return snapshotCopy_; 15854 } 15855 15856 // optional uint32 createdListSize = 6; 15857 public static final int CREATEDLISTSIZE_FIELD_NUMBER = 6; 15858 private int createdListSize_; 15859 /** 15860 * <code>optional uint32 createdListSize = 6;</code> 15861 */ 15862 public boolean hasCreatedListSize() { 15863 return ((bitField0_ & 0x00000020) == 0x00000020); 15864 } 15865 /** 15866 * <code>optional uint32 createdListSize = 6;</code> 15867 */ 15868 public int getCreatedListSize() { 15869 return createdListSize_; 15870 } 15871 15872 // repeated uint64 deletedINode = 7 [packed = true]; 15873 public static final int DELETEDINODE_FIELD_NUMBER = 7; 15874 private java.util.List<java.lang.Long> deletedINode_; 15875 /** 15876 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15877 * 15878 * <pre> 15879 * id of deleted inodes 15880 * </pre> 15881 */ 15882 public java.util.List<java.lang.Long> 15883 getDeletedINodeList() { 15884 return deletedINode_; 15885 } 15886 /** 15887 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15888 * 15889 * <pre> 15890 * id of deleted inodes 15891 * </pre> 15892 */ 15893 public int getDeletedINodeCount() { 15894 return deletedINode_.size(); 15895 } 15896 /** 15897 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 15898 * 15899 * <pre> 15900 * id of deleted inodes 15901 * </pre> 15902 */ 15903 public long getDeletedINode(int index) { 15904 return deletedINode_.get(index); 15905 } 15906 private int deletedINodeMemoizedSerializedSize = -1; 15907 15908 // repeated uint32 deletedINodeRef = 8 [packed = true]; 15909 public static final int DELETEDINODEREF_FIELD_NUMBER = 8; 15910 private java.util.List<java.lang.Integer> deletedINodeRef_; 15911 /** 15912 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15913 * 15914 * <pre> 15915 * id of reference nodes in the deleted list 15916 * </pre> 15917 */ 15918 public java.util.List<java.lang.Integer> 15919 getDeletedINodeRefList() { 15920 return deletedINodeRef_; 15921 } 15922 /** 15923 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15924 * 15925 * <pre> 15926 * id of reference nodes in the deleted list 15927 * </pre> 15928 */ 15929 public int getDeletedINodeRefCount() { 15930 return deletedINodeRef_.size(); 15931 } 15932 /** 15933 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 15934 * 15935 * <pre> 15936 * id of reference nodes in the deleted list 15937 * </pre> 15938 */ 15939 public int getDeletedINodeRef(int index) { 15940 return deletedINodeRef_.get(index); 15941 } 15942 private int deletedINodeRefMemoizedSerializedSize = -1; 15943 15944 private void initFields() { 15945 snapshotId_ = 0; 15946 childrenSize_ = 0; 15947 isSnapshotRoot_ = false; 15948 name_ = com.google.protobuf.ByteString.EMPTY; 15949 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 15950 createdListSize_ = 0; 15951 deletedINode_ = java.util.Collections.emptyList(); 15952 deletedINodeRef_ = java.util.Collections.emptyList(); 15953 } 15954 private byte memoizedIsInitialized = -1; 15955 public final boolean isInitialized() { 15956 byte isInitialized = memoizedIsInitialized; 15957 if (isInitialized != -1) return isInitialized == 1; 15958 15959 if (hasSnapshotCopy()) { 15960 if (!getSnapshotCopy().isInitialized()) { 15961 memoizedIsInitialized = 0; 15962 return false; 15963 } 15964 } 15965 memoizedIsInitialized = 1; 15966 return true; 15967 } 15968 15969 public void writeTo(com.google.protobuf.CodedOutputStream output) 15970 throws java.io.IOException { 15971 getSerializedSize(); 15972 if (((bitField0_ & 0x00000001) == 0x00000001)) { 15973 output.writeUInt32(1, snapshotId_); 15974 } 15975 if (((bitField0_ & 0x00000002) == 0x00000002)) { 15976 output.writeUInt32(2, childrenSize_); 15977 } 15978 if (((bitField0_ & 0x00000004) == 0x00000004)) { 15979 output.writeBool(3, isSnapshotRoot_); 15980 } 15981 if (((bitField0_ & 0x00000008) == 0x00000008)) { 15982 output.writeBytes(4, name_); 15983 } 15984 if (((bitField0_ & 0x00000010) == 0x00000010)) { 15985 output.writeMessage(5, snapshotCopy_); 15986 } 15987 if (((bitField0_ & 0x00000020) == 0x00000020)) { 15988 output.writeUInt32(6, createdListSize_); 15989 } 15990 if (getDeletedINodeList().size() > 0) { 15991 output.writeRawVarint32(58); 15992 output.writeRawVarint32(deletedINodeMemoizedSerializedSize); 15993 } 15994 for (int i = 0; i < deletedINode_.size(); i++) { 15995 output.writeUInt64NoTag(deletedINode_.get(i)); 15996 } 15997 if (getDeletedINodeRefList().size() > 0) { 15998 output.writeRawVarint32(66); 15999 output.writeRawVarint32(deletedINodeRefMemoizedSerializedSize); 16000 } 16001 for (int i = 0; i < deletedINodeRef_.size(); i++) { 16002 output.writeUInt32NoTag(deletedINodeRef_.get(i)); 16003 } 16004 getUnknownFields().writeTo(output); 16005 } 16006 16007 private int memoizedSerializedSize = -1; 16008 public int getSerializedSize() { 16009 int size = memoizedSerializedSize; 16010 if (size != -1) return size; 16011 16012 size = 0; 16013 if (((bitField0_ & 0x00000001) == 0x00000001)) { 16014 size += com.google.protobuf.CodedOutputStream 16015 .computeUInt32Size(1, snapshotId_); 16016 } 16017 if (((bitField0_ & 0x00000002) == 0x00000002)) { 16018 size += com.google.protobuf.CodedOutputStream 16019 .computeUInt32Size(2, childrenSize_); 16020 } 16021 if (((bitField0_ & 0x00000004) == 0x00000004)) { 16022 size += com.google.protobuf.CodedOutputStream 16023 .computeBoolSize(3, isSnapshotRoot_); 16024 } 16025 if (((bitField0_ & 0x00000008) == 0x00000008)) { 16026 size += com.google.protobuf.CodedOutputStream 16027 .computeBytesSize(4, name_); 16028 } 16029 if (((bitField0_ & 0x00000010) == 0x00000010)) { 16030 size += com.google.protobuf.CodedOutputStream 16031 .computeMessageSize(5, snapshotCopy_); 16032 } 16033 if (((bitField0_ & 0x00000020) == 0x00000020)) { 16034 size += com.google.protobuf.CodedOutputStream 16035 .computeUInt32Size(6, createdListSize_); 16036 } 16037 { 16038 int dataSize = 0; 16039 for (int i = 0; i < deletedINode_.size(); i++) { 16040 dataSize += com.google.protobuf.CodedOutputStream 16041 .computeUInt64SizeNoTag(deletedINode_.get(i)); 16042 } 16043 size += dataSize; 16044 if (!getDeletedINodeList().isEmpty()) { 16045 size += 1; 16046 size += com.google.protobuf.CodedOutputStream 16047 .computeInt32SizeNoTag(dataSize); 16048 } 16049 deletedINodeMemoizedSerializedSize = dataSize; 16050 } 16051 { 16052 int dataSize = 0; 16053 for (int i = 0; i < deletedINodeRef_.size(); i++) { 16054 dataSize += com.google.protobuf.CodedOutputStream 16055 .computeUInt32SizeNoTag(deletedINodeRef_.get(i)); 16056 } 16057 size += dataSize; 16058 if (!getDeletedINodeRefList().isEmpty()) { 16059 size += 1; 16060 size += com.google.protobuf.CodedOutputStream 16061 .computeInt32SizeNoTag(dataSize); 16062 } 16063 deletedINodeRefMemoizedSerializedSize = dataSize; 16064 } 16065 size += getUnknownFields().getSerializedSize(); 16066 memoizedSerializedSize = size; 16067 return size; 16068 } 16069 16070 private static final long serialVersionUID = 0L; 16071 @java.lang.Override 16072 protected java.lang.Object writeReplace() 16073 throws java.io.ObjectStreamException { 16074 return super.writeReplace(); 16075 } 16076 16077 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16078 com.google.protobuf.ByteString data) 16079 throws com.google.protobuf.InvalidProtocolBufferException { 16080 return PARSER.parseFrom(data); 16081 } 16082 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16083 com.google.protobuf.ByteString data, 16084 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16085 throws com.google.protobuf.InvalidProtocolBufferException { 16086 return PARSER.parseFrom(data, extensionRegistry); 16087 } 16088 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom(byte[] data) 16089 throws com.google.protobuf.InvalidProtocolBufferException { 16090 return PARSER.parseFrom(data); 16091 } 16092 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16093 byte[] data, 16094 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16095 throws com.google.protobuf.InvalidProtocolBufferException { 16096 return PARSER.parseFrom(data, extensionRegistry); 16097 } 16098 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom(java.io.InputStream input) 16099 throws java.io.IOException { 16100 return PARSER.parseFrom(input); 16101 } 16102 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16103 java.io.InputStream input, 16104 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16105 throws java.io.IOException { 16106 return PARSER.parseFrom(input, extensionRegistry); 16107 } 16108 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseDelimitedFrom(java.io.InputStream input) 16109 throws java.io.IOException { 16110 return PARSER.parseDelimitedFrom(input); 16111 } 16112 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseDelimitedFrom( 16113 java.io.InputStream input, 16114 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16115 throws java.io.IOException { 16116 return PARSER.parseDelimitedFrom(input, extensionRegistry); 16117 } 16118 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16119 com.google.protobuf.CodedInputStream input) 16120 throws java.io.IOException { 16121 return PARSER.parseFrom(input); 16122 } 16123 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parseFrom( 16124 com.google.protobuf.CodedInputStream input, 16125 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16126 throws java.io.IOException { 16127 return PARSER.parseFrom(input, extensionRegistry); 16128 } 16129 16130 public static Builder newBuilder() { return Builder.create(); } 16131 public Builder newBuilderForType() { return newBuilder(); } 16132 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff prototype) { 16133 return newBuilder().mergeFrom(prototype); 16134 } 16135 public Builder toBuilder() { return newBuilder(this); } 16136 16137 @java.lang.Override 16138 protected Builder newBuilderForType( 16139 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16140 Builder builder = new Builder(parent); 16141 return builder; 16142 } 16143 /** 16144 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff} 16145 */ 16146 public static final class Builder extends 16147 com.google.protobuf.GeneratedMessage.Builder<Builder> 16148 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiffOrBuilder { 16149 public static final com.google.protobuf.Descriptors.Descriptor 16150 getDescriptor() { 16151 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 16152 } 16153 16154 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16155 internalGetFieldAccessorTable() { 16156 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable 16157 .ensureFieldAccessorsInitialized( 16158 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.Builder.class); 16159 } 16160 16161 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.newBuilder() 16162 private Builder() { 16163 maybeForceBuilderInitialization(); 16164 } 16165 16166 private Builder( 16167 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 16168 super(parent); 16169 maybeForceBuilderInitialization(); 16170 } 16171 private void maybeForceBuilderInitialization() { 16172 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 16173 getSnapshotCopyFieldBuilder(); 16174 } 16175 } 16176 private static Builder create() { 16177 return new Builder(); 16178 } 16179 16180 public Builder clear() { 16181 super.clear(); 16182 snapshotId_ = 0; 16183 bitField0_ = (bitField0_ & ~0x00000001); 16184 childrenSize_ = 0; 16185 bitField0_ = (bitField0_ & ~0x00000002); 16186 isSnapshotRoot_ = false; 16187 bitField0_ = (bitField0_ & ~0x00000004); 16188 name_ = com.google.protobuf.ByteString.EMPTY; 16189 bitField0_ = (bitField0_ & ~0x00000008); 16190 if (snapshotCopyBuilder_ == null) { 16191 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 16192 } else { 16193 snapshotCopyBuilder_.clear(); 16194 } 16195 bitField0_ = (bitField0_ & ~0x00000010); 16196 createdListSize_ = 0; 16197 bitField0_ = (bitField0_ & ~0x00000020); 16198 deletedINode_ = java.util.Collections.emptyList(); 16199 bitField0_ = (bitField0_ & ~0x00000040); 16200 deletedINodeRef_ = java.util.Collections.emptyList(); 16201 bitField0_ = (bitField0_ & ~0x00000080); 16202 return this; 16203 } 16204 16205 public Builder clone() { 16206 return create().mergeFrom(buildPartial()); 16207 } 16208 16209 public com.google.protobuf.Descriptors.Descriptor 16210 getDescriptorForType() { 16211 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 16212 } 16213 16214 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff getDefaultInstanceForType() { 16215 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.getDefaultInstance(); 16216 } 16217 16218 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff build() { 16219 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff result = buildPartial(); 16220 if (!result.isInitialized()) { 16221 throw newUninitializedMessageException(result); 16222 } 16223 return result; 16224 } 16225 16226 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff buildPartial() { 16227 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff(this); 16228 int from_bitField0_ = bitField0_; 16229 int to_bitField0_ = 0; 16230 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 16231 to_bitField0_ |= 0x00000001; 16232 } 16233 result.snapshotId_ = snapshotId_; 16234 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 16235 to_bitField0_ |= 0x00000002; 16236 } 16237 result.childrenSize_ = childrenSize_; 16238 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 16239 to_bitField0_ |= 0x00000004; 16240 } 16241 result.isSnapshotRoot_ = isSnapshotRoot_; 16242 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 16243 to_bitField0_ |= 0x00000008; 16244 } 16245 result.name_ = name_; 16246 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 16247 to_bitField0_ |= 0x00000010; 16248 } 16249 if (snapshotCopyBuilder_ == null) { 16250 result.snapshotCopy_ = snapshotCopy_; 16251 } else { 16252 result.snapshotCopy_ = snapshotCopyBuilder_.build(); 16253 } 16254 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 16255 to_bitField0_ |= 0x00000020; 16256 } 16257 result.createdListSize_ = createdListSize_; 16258 if (((bitField0_ & 0x00000040) == 0x00000040)) { 16259 deletedINode_ = java.util.Collections.unmodifiableList(deletedINode_); 16260 bitField0_ = (bitField0_ & ~0x00000040); 16261 } 16262 result.deletedINode_ = deletedINode_; 16263 if (((bitField0_ & 0x00000080) == 0x00000080)) { 16264 deletedINodeRef_ = java.util.Collections.unmodifiableList(deletedINodeRef_); 16265 bitField0_ = (bitField0_ & ~0x00000080); 16266 } 16267 result.deletedINodeRef_ = deletedINodeRef_; 16268 result.bitField0_ = to_bitField0_; 16269 onBuilt(); 16270 return result; 16271 } 16272 16273 public Builder mergeFrom(com.google.protobuf.Message other) { 16274 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff) { 16275 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff)other); 16276 } else { 16277 super.mergeFrom(other); 16278 return this; 16279 } 16280 } 16281 16282 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff other) { 16283 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff.getDefaultInstance()) return this; 16284 if (other.hasSnapshotId()) { 16285 setSnapshotId(other.getSnapshotId()); 16286 } 16287 if (other.hasChildrenSize()) { 16288 setChildrenSize(other.getChildrenSize()); 16289 } 16290 if (other.hasIsSnapshotRoot()) { 16291 setIsSnapshotRoot(other.getIsSnapshotRoot()); 16292 } 16293 if (other.hasName()) { 16294 setName(other.getName()); 16295 } 16296 if (other.hasSnapshotCopy()) { 16297 mergeSnapshotCopy(other.getSnapshotCopy()); 16298 } 16299 if (other.hasCreatedListSize()) { 16300 setCreatedListSize(other.getCreatedListSize()); 16301 } 16302 if (!other.deletedINode_.isEmpty()) { 16303 if (deletedINode_.isEmpty()) { 16304 deletedINode_ = other.deletedINode_; 16305 bitField0_ = (bitField0_ & ~0x00000040); 16306 } else { 16307 ensureDeletedINodeIsMutable(); 16308 deletedINode_.addAll(other.deletedINode_); 16309 } 16310 onChanged(); 16311 } 16312 if (!other.deletedINodeRef_.isEmpty()) { 16313 if (deletedINodeRef_.isEmpty()) { 16314 deletedINodeRef_ = other.deletedINodeRef_; 16315 bitField0_ = (bitField0_ & ~0x00000080); 16316 } else { 16317 ensureDeletedINodeRefIsMutable(); 16318 deletedINodeRef_.addAll(other.deletedINodeRef_); 16319 } 16320 onChanged(); 16321 } 16322 this.mergeUnknownFields(other.getUnknownFields()); 16323 return this; 16324 } 16325 16326 public final boolean isInitialized() { 16327 if (hasSnapshotCopy()) { 16328 if (!getSnapshotCopy().isInitialized()) { 16329 16330 return false; 16331 } 16332 } 16333 return true; 16334 } 16335 16336 public Builder mergeFrom( 16337 com.google.protobuf.CodedInputStream input, 16338 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16339 throws java.io.IOException { 16340 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff parsedMessage = null; 16341 try { 16342 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 16343 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16344 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DirectoryDiff) e.getUnfinishedMessage(); 16345 throw e; 16346 } finally { 16347 if (parsedMessage != null) { 16348 mergeFrom(parsedMessage); 16349 } 16350 } 16351 return this; 16352 } 16353 private int bitField0_; 16354 16355 // optional uint32 snapshotId = 1; 16356 private int snapshotId_ ; 16357 /** 16358 * <code>optional uint32 snapshotId = 1;</code> 16359 */ 16360 public boolean hasSnapshotId() { 16361 return ((bitField0_ & 0x00000001) == 0x00000001); 16362 } 16363 /** 16364 * <code>optional uint32 snapshotId = 1;</code> 16365 */ 16366 public int getSnapshotId() { 16367 return snapshotId_; 16368 } 16369 /** 16370 * <code>optional uint32 snapshotId = 1;</code> 16371 */ 16372 public Builder setSnapshotId(int value) { 16373 bitField0_ |= 0x00000001; 16374 snapshotId_ = value; 16375 onChanged(); 16376 return this; 16377 } 16378 /** 16379 * <code>optional uint32 snapshotId = 1;</code> 16380 */ 16381 public Builder clearSnapshotId() { 16382 bitField0_ = (bitField0_ & ~0x00000001); 16383 snapshotId_ = 0; 16384 onChanged(); 16385 return this; 16386 } 16387 16388 // optional uint32 childrenSize = 2; 16389 private int childrenSize_ ; 16390 /** 16391 * <code>optional uint32 childrenSize = 2;</code> 16392 */ 16393 public boolean hasChildrenSize() { 16394 return ((bitField0_ & 0x00000002) == 0x00000002); 16395 } 16396 /** 16397 * <code>optional uint32 childrenSize = 2;</code> 16398 */ 16399 public int getChildrenSize() { 16400 return childrenSize_; 16401 } 16402 /** 16403 * <code>optional uint32 childrenSize = 2;</code> 16404 */ 16405 public Builder setChildrenSize(int value) { 16406 bitField0_ |= 0x00000002; 16407 childrenSize_ = value; 16408 onChanged(); 16409 return this; 16410 } 16411 /** 16412 * <code>optional uint32 childrenSize = 2;</code> 16413 */ 16414 public Builder clearChildrenSize() { 16415 bitField0_ = (bitField0_ & ~0x00000002); 16416 childrenSize_ = 0; 16417 onChanged(); 16418 return this; 16419 } 16420 16421 // optional bool isSnapshotRoot = 3; 16422 private boolean isSnapshotRoot_ ; 16423 /** 16424 * <code>optional bool isSnapshotRoot = 3;</code> 16425 */ 16426 public boolean hasIsSnapshotRoot() { 16427 return ((bitField0_ & 0x00000004) == 0x00000004); 16428 } 16429 /** 16430 * <code>optional bool isSnapshotRoot = 3;</code> 16431 */ 16432 public boolean getIsSnapshotRoot() { 16433 return isSnapshotRoot_; 16434 } 16435 /** 16436 * <code>optional bool isSnapshotRoot = 3;</code> 16437 */ 16438 public Builder setIsSnapshotRoot(boolean value) { 16439 bitField0_ |= 0x00000004; 16440 isSnapshotRoot_ = value; 16441 onChanged(); 16442 return this; 16443 } 16444 /** 16445 * <code>optional bool isSnapshotRoot = 3;</code> 16446 */ 16447 public Builder clearIsSnapshotRoot() { 16448 bitField0_ = (bitField0_ & ~0x00000004); 16449 isSnapshotRoot_ = false; 16450 onChanged(); 16451 return this; 16452 } 16453 16454 // optional bytes name = 4; 16455 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 16456 /** 16457 * <code>optional bytes name = 4;</code> 16458 */ 16459 public boolean hasName() { 16460 return ((bitField0_ & 0x00000008) == 0x00000008); 16461 } 16462 /** 16463 * <code>optional bytes name = 4;</code> 16464 */ 16465 public com.google.protobuf.ByteString getName() { 16466 return name_; 16467 } 16468 /** 16469 * <code>optional bytes name = 4;</code> 16470 */ 16471 public Builder setName(com.google.protobuf.ByteString value) { 16472 if (value == null) { 16473 throw new NullPointerException(); 16474 } 16475 bitField0_ |= 0x00000008; 16476 name_ = value; 16477 onChanged(); 16478 return this; 16479 } 16480 /** 16481 * <code>optional bytes name = 4;</code> 16482 */ 16483 public Builder clearName() { 16484 bitField0_ = (bitField0_ & ~0x00000008); 16485 name_ = getDefaultInstance().getName(); 16486 onChanged(); 16487 return this; 16488 } 16489 16490 // optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5; 16491 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 16492 private com.google.protobuf.SingleFieldBuilder< 16493 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> snapshotCopyBuilder_; 16494 /** 16495 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16496 */ 16497 public boolean hasSnapshotCopy() { 16498 return ((bitField0_ & 0x00000010) == 0x00000010); 16499 } 16500 /** 16501 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16502 */ 16503 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory getSnapshotCopy() { 16504 if (snapshotCopyBuilder_ == null) { 16505 return snapshotCopy_; 16506 } else { 16507 return snapshotCopyBuilder_.getMessage(); 16508 } 16509 } 16510 /** 16511 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16512 */ 16513 public Builder setSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 16514 if (snapshotCopyBuilder_ == null) { 16515 if (value == null) { 16516 throw new NullPointerException(); 16517 } 16518 snapshotCopy_ = value; 16519 onChanged(); 16520 } else { 16521 snapshotCopyBuilder_.setMessage(value); 16522 } 16523 bitField0_ |= 0x00000010; 16524 return this; 16525 } 16526 /** 16527 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16528 */ 16529 public Builder setSnapshotCopy( 16530 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder builderForValue) { 16531 if (snapshotCopyBuilder_ == null) { 16532 snapshotCopy_ = builderForValue.build(); 16533 onChanged(); 16534 } else { 16535 snapshotCopyBuilder_.setMessage(builderForValue.build()); 16536 } 16537 bitField0_ |= 0x00000010; 16538 return this; 16539 } 16540 /** 16541 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16542 */ 16543 public Builder mergeSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory value) { 16544 if (snapshotCopyBuilder_ == null) { 16545 if (((bitField0_ & 0x00000010) == 0x00000010) && 16546 snapshotCopy_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance()) { 16547 snapshotCopy_ = 16548 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.newBuilder(snapshotCopy_).mergeFrom(value).buildPartial(); 16549 } else { 16550 snapshotCopy_ = value; 16551 } 16552 onChanged(); 16553 } else { 16554 snapshotCopyBuilder_.mergeFrom(value); 16555 } 16556 bitField0_ |= 0x00000010; 16557 return this; 16558 } 16559 /** 16560 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16561 */ 16562 public Builder clearSnapshotCopy() { 16563 if (snapshotCopyBuilder_ == null) { 16564 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.getDefaultInstance(); 16565 onChanged(); 16566 } else { 16567 snapshotCopyBuilder_.clear(); 16568 } 16569 bitField0_ = (bitField0_ & ~0x00000010); 16570 return this; 16571 } 16572 /** 16573 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16574 */ 16575 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder getSnapshotCopyBuilder() { 16576 bitField0_ |= 0x00000010; 16577 onChanged(); 16578 return getSnapshotCopyFieldBuilder().getBuilder(); 16579 } 16580 /** 16581 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16582 */ 16583 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder getSnapshotCopyOrBuilder() { 16584 if (snapshotCopyBuilder_ != null) { 16585 return snapshotCopyBuilder_.getMessageOrBuilder(); 16586 } else { 16587 return snapshotCopy_; 16588 } 16589 } 16590 /** 16591 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeDirectory snapshotCopy = 5;</code> 16592 */ 16593 private com.google.protobuf.SingleFieldBuilder< 16594 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder> 16595 getSnapshotCopyFieldBuilder() { 16596 if (snapshotCopyBuilder_ == null) { 16597 snapshotCopyBuilder_ = new com.google.protobuf.SingleFieldBuilder< 16598 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectory.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeDirectoryOrBuilder>( 16599 snapshotCopy_, 16600 getParentForChildren(), 16601 isClean()); 16602 snapshotCopy_ = null; 16603 } 16604 return snapshotCopyBuilder_; 16605 } 16606 16607 // optional uint32 createdListSize = 6; 16608 private int createdListSize_ ; 16609 /** 16610 * <code>optional uint32 createdListSize = 6;</code> 16611 */ 16612 public boolean hasCreatedListSize() { 16613 return ((bitField0_ & 0x00000020) == 0x00000020); 16614 } 16615 /** 16616 * <code>optional uint32 createdListSize = 6;</code> 16617 */ 16618 public int getCreatedListSize() { 16619 return createdListSize_; 16620 } 16621 /** 16622 * <code>optional uint32 createdListSize = 6;</code> 16623 */ 16624 public Builder setCreatedListSize(int value) { 16625 bitField0_ |= 0x00000020; 16626 createdListSize_ = value; 16627 onChanged(); 16628 return this; 16629 } 16630 /** 16631 * <code>optional uint32 createdListSize = 6;</code> 16632 */ 16633 public Builder clearCreatedListSize() { 16634 bitField0_ = (bitField0_ & ~0x00000020); 16635 createdListSize_ = 0; 16636 onChanged(); 16637 return this; 16638 } 16639 16640 // repeated uint64 deletedINode = 7 [packed = true]; 16641 private java.util.List<java.lang.Long> deletedINode_ = java.util.Collections.emptyList(); 16642 private void ensureDeletedINodeIsMutable() { 16643 if (!((bitField0_ & 0x00000040) == 0x00000040)) { 16644 deletedINode_ = new java.util.ArrayList<java.lang.Long>(deletedINode_); 16645 bitField0_ |= 0x00000040; 16646 } 16647 } 16648 /** 16649 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16650 * 16651 * <pre> 16652 * id of deleted inodes 16653 * </pre> 16654 */ 16655 public java.util.List<java.lang.Long> 16656 getDeletedINodeList() { 16657 return java.util.Collections.unmodifiableList(deletedINode_); 16658 } 16659 /** 16660 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16661 * 16662 * <pre> 16663 * id of deleted inodes 16664 * </pre> 16665 */ 16666 public int getDeletedINodeCount() { 16667 return deletedINode_.size(); 16668 } 16669 /** 16670 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16671 * 16672 * <pre> 16673 * id of deleted inodes 16674 * </pre> 16675 */ 16676 public long getDeletedINode(int index) { 16677 return deletedINode_.get(index); 16678 } 16679 /** 16680 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16681 * 16682 * <pre> 16683 * id of deleted inodes 16684 * </pre> 16685 */ 16686 public Builder setDeletedINode( 16687 int index, long value) { 16688 ensureDeletedINodeIsMutable(); 16689 deletedINode_.set(index, value); 16690 onChanged(); 16691 return this; 16692 } 16693 /** 16694 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16695 * 16696 * <pre> 16697 * id of deleted inodes 16698 * </pre> 16699 */ 16700 public Builder addDeletedINode(long value) { 16701 ensureDeletedINodeIsMutable(); 16702 deletedINode_.add(value); 16703 onChanged(); 16704 return this; 16705 } 16706 /** 16707 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16708 * 16709 * <pre> 16710 * id of deleted inodes 16711 * </pre> 16712 */ 16713 public Builder addAllDeletedINode( 16714 java.lang.Iterable<? extends java.lang.Long> values) { 16715 ensureDeletedINodeIsMutable(); 16716 super.addAll(values, deletedINode_); 16717 onChanged(); 16718 return this; 16719 } 16720 /** 16721 * <code>repeated uint64 deletedINode = 7 [packed = true];</code> 16722 * 16723 * <pre> 16724 * id of deleted inodes 16725 * </pre> 16726 */ 16727 public Builder clearDeletedINode() { 16728 deletedINode_ = java.util.Collections.emptyList(); 16729 bitField0_ = (bitField0_ & ~0x00000040); 16730 onChanged(); 16731 return this; 16732 } 16733 16734 // repeated uint32 deletedINodeRef = 8 [packed = true]; 16735 private java.util.List<java.lang.Integer> deletedINodeRef_ = java.util.Collections.emptyList(); 16736 private void ensureDeletedINodeRefIsMutable() { 16737 if (!((bitField0_ & 0x00000080) == 0x00000080)) { 16738 deletedINodeRef_ = new java.util.ArrayList<java.lang.Integer>(deletedINodeRef_); 16739 bitField0_ |= 0x00000080; 16740 } 16741 } 16742 /** 16743 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16744 * 16745 * <pre> 16746 * id of reference nodes in the deleted list 16747 * </pre> 16748 */ 16749 public java.util.List<java.lang.Integer> 16750 getDeletedINodeRefList() { 16751 return java.util.Collections.unmodifiableList(deletedINodeRef_); 16752 } 16753 /** 16754 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16755 * 16756 * <pre> 16757 * id of reference nodes in the deleted list 16758 * </pre> 16759 */ 16760 public int getDeletedINodeRefCount() { 16761 return deletedINodeRef_.size(); 16762 } 16763 /** 16764 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16765 * 16766 * <pre> 16767 * id of reference nodes in the deleted list 16768 * </pre> 16769 */ 16770 public int getDeletedINodeRef(int index) { 16771 return deletedINodeRef_.get(index); 16772 } 16773 /** 16774 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16775 * 16776 * <pre> 16777 * id of reference nodes in the deleted list 16778 * </pre> 16779 */ 16780 public Builder setDeletedINodeRef( 16781 int index, int value) { 16782 ensureDeletedINodeRefIsMutable(); 16783 deletedINodeRef_.set(index, value); 16784 onChanged(); 16785 return this; 16786 } 16787 /** 16788 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16789 * 16790 * <pre> 16791 * id of reference nodes in the deleted list 16792 * </pre> 16793 */ 16794 public Builder addDeletedINodeRef(int value) { 16795 ensureDeletedINodeRefIsMutable(); 16796 deletedINodeRef_.add(value); 16797 onChanged(); 16798 return this; 16799 } 16800 /** 16801 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16802 * 16803 * <pre> 16804 * id of reference nodes in the deleted list 16805 * </pre> 16806 */ 16807 public Builder addAllDeletedINodeRef( 16808 java.lang.Iterable<? extends java.lang.Integer> values) { 16809 ensureDeletedINodeRefIsMutable(); 16810 super.addAll(values, deletedINodeRef_); 16811 onChanged(); 16812 return this; 16813 } 16814 /** 16815 * <code>repeated uint32 deletedINodeRef = 8 [packed = true];</code> 16816 * 16817 * <pre> 16818 * id of reference nodes in the deleted list 16819 * </pre> 16820 */ 16821 public Builder clearDeletedINodeRef() { 16822 deletedINodeRef_ = java.util.Collections.emptyList(); 16823 bitField0_ = (bitField0_ & ~0x00000080); 16824 onChanged(); 16825 return this; 16826 } 16827 16828 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff) 16829 } 16830 16831 static { 16832 defaultInstance = new DirectoryDiff(true); 16833 defaultInstance.initFields(); 16834 } 16835 16836 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DirectoryDiff) 16837 } 16838 16839 public interface FileDiffOrBuilder 16840 extends com.google.protobuf.MessageOrBuilder { 16841 16842 // optional uint32 snapshotId = 1; 16843 /** 16844 * <code>optional uint32 snapshotId = 1;</code> 16845 */ 16846 boolean hasSnapshotId(); 16847 /** 16848 * <code>optional uint32 snapshotId = 1;</code> 16849 */ 16850 int getSnapshotId(); 16851 16852 // optional uint64 fileSize = 2; 16853 /** 16854 * <code>optional uint64 fileSize = 2;</code> 16855 */ 16856 boolean hasFileSize(); 16857 /** 16858 * <code>optional uint64 fileSize = 2;</code> 16859 */ 16860 long getFileSize(); 16861 16862 // optional bytes name = 3; 16863 /** 16864 * <code>optional bytes name = 3;</code> 16865 */ 16866 boolean hasName(); 16867 /** 16868 * <code>optional bytes name = 3;</code> 16869 */ 16870 com.google.protobuf.ByteString getName(); 16871 16872 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 16873 /** 16874 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 16875 */ 16876 boolean hasSnapshotCopy(); 16877 /** 16878 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 16879 */ 16880 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy(); 16881 /** 16882 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 16883 */ 16884 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder(); 16885 } 16886 /** 16887 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff} 16888 */ 16889 public static final class FileDiff extends 16890 com.google.protobuf.GeneratedMessage 16891 implements FileDiffOrBuilder { 16892 // Use FileDiff.newBuilder() to construct. 16893 private FileDiff(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 16894 super(builder); 16895 this.unknownFields = builder.getUnknownFields(); 16896 } 16897 private FileDiff(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 16898 16899 private static final FileDiff defaultInstance; 16900 public static FileDiff getDefaultInstance() { 16901 return defaultInstance; 16902 } 16903 16904 public FileDiff getDefaultInstanceForType() { 16905 return defaultInstance; 16906 } 16907 16908 private final com.google.protobuf.UnknownFieldSet unknownFields; 16909 @java.lang.Override 16910 public final com.google.protobuf.UnknownFieldSet 16911 getUnknownFields() { 16912 return this.unknownFields; 16913 } 16914 private FileDiff( 16915 com.google.protobuf.CodedInputStream input, 16916 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16917 throws com.google.protobuf.InvalidProtocolBufferException { 16918 initFields(); 16919 int mutable_bitField0_ = 0; 16920 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 16921 com.google.protobuf.UnknownFieldSet.newBuilder(); 16922 try { 16923 boolean done = false; 16924 while (!done) { 16925 int tag = input.readTag(); 16926 switch (tag) { 16927 case 0: 16928 done = true; 16929 break; 16930 default: { 16931 if (!parseUnknownField(input, unknownFields, 16932 extensionRegistry, tag)) { 16933 done = true; 16934 } 16935 break; 16936 } 16937 case 8: { 16938 bitField0_ |= 0x00000001; 16939 snapshotId_ = input.readUInt32(); 16940 break; 16941 } 16942 case 16: { 16943 bitField0_ |= 0x00000002; 16944 fileSize_ = input.readUInt64(); 16945 break; 16946 } 16947 case 26: { 16948 bitField0_ |= 0x00000004; 16949 name_ = input.readBytes(); 16950 break; 16951 } 16952 case 34: { 16953 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder subBuilder = null; 16954 if (((bitField0_ & 0x00000008) == 0x00000008)) { 16955 subBuilder = snapshotCopy_.toBuilder(); 16956 } 16957 snapshotCopy_ = input.readMessage(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.PARSER, extensionRegistry); 16958 if (subBuilder != null) { 16959 subBuilder.mergeFrom(snapshotCopy_); 16960 snapshotCopy_ = subBuilder.buildPartial(); 16961 } 16962 bitField0_ |= 0x00000008; 16963 break; 16964 } 16965 } 16966 } 16967 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 16968 throw e.setUnfinishedMessage(this); 16969 } catch (java.io.IOException e) { 16970 throw new com.google.protobuf.InvalidProtocolBufferException( 16971 e.getMessage()).setUnfinishedMessage(this); 16972 } finally { 16973 this.unknownFields = unknownFields.build(); 16974 makeExtensionsImmutable(); 16975 } 16976 } 16977 public static final com.google.protobuf.Descriptors.Descriptor 16978 getDescriptor() { 16979 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 16980 } 16981 16982 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 16983 internalGetFieldAccessorTable() { 16984 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable 16985 .ensureFieldAccessorsInitialized( 16986 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.Builder.class); 16987 } 16988 16989 public static com.google.protobuf.Parser<FileDiff> PARSER = 16990 new com.google.protobuf.AbstractParser<FileDiff>() { 16991 public FileDiff parsePartialFrom( 16992 com.google.protobuf.CodedInputStream input, 16993 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 16994 throws com.google.protobuf.InvalidProtocolBufferException { 16995 return new FileDiff(input, extensionRegistry); 16996 } 16997 }; 16998 16999 @java.lang.Override 17000 public com.google.protobuf.Parser<FileDiff> getParserForType() { 17001 return PARSER; 17002 } 17003 17004 private int bitField0_; 17005 // optional uint32 snapshotId = 1; 17006 public static final int SNAPSHOTID_FIELD_NUMBER = 1; 17007 private int snapshotId_; 17008 /** 17009 * <code>optional uint32 snapshotId = 1;</code> 17010 */ 17011 public boolean hasSnapshotId() { 17012 return ((bitField0_ & 0x00000001) == 0x00000001); 17013 } 17014 /** 17015 * <code>optional uint32 snapshotId = 1;</code> 17016 */ 17017 public int getSnapshotId() { 17018 return snapshotId_; 17019 } 17020 17021 // optional uint64 fileSize = 2; 17022 public static final int FILESIZE_FIELD_NUMBER = 2; 17023 private long fileSize_; 17024 /** 17025 * <code>optional uint64 fileSize = 2;</code> 17026 */ 17027 public boolean hasFileSize() { 17028 return ((bitField0_ & 0x00000002) == 0x00000002); 17029 } 17030 /** 17031 * <code>optional uint64 fileSize = 2;</code> 17032 */ 17033 public long getFileSize() { 17034 return fileSize_; 17035 } 17036 17037 // optional bytes name = 3; 17038 public static final int NAME_FIELD_NUMBER = 3; 17039 private com.google.protobuf.ByteString name_; 17040 /** 17041 * <code>optional bytes name = 3;</code> 17042 */ 17043 public boolean hasName() { 17044 return ((bitField0_ & 0x00000004) == 0x00000004); 17045 } 17046 /** 17047 * <code>optional bytes name = 3;</code> 17048 */ 17049 public com.google.protobuf.ByteString getName() { 17050 return name_; 17051 } 17052 17053 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 17054 public static final int SNAPSHOTCOPY_FIELD_NUMBER = 4; 17055 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile snapshotCopy_; 17056 /** 17057 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17058 */ 17059 public boolean hasSnapshotCopy() { 17060 return ((bitField0_ & 0x00000008) == 0x00000008); 17061 } 17062 /** 17063 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17064 */ 17065 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy() { 17066 return snapshotCopy_; 17067 } 17068 /** 17069 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17070 */ 17071 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder() { 17072 return snapshotCopy_; 17073 } 17074 17075 private void initFields() { 17076 snapshotId_ = 0; 17077 fileSize_ = 0L; 17078 name_ = com.google.protobuf.ByteString.EMPTY; 17079 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 17080 } 17081 private byte memoizedIsInitialized = -1; 17082 public final boolean isInitialized() { 17083 byte isInitialized = memoizedIsInitialized; 17084 if (isInitialized != -1) return isInitialized == 1; 17085 17086 if (hasSnapshotCopy()) { 17087 if (!getSnapshotCopy().isInitialized()) { 17088 memoizedIsInitialized = 0; 17089 return false; 17090 } 17091 } 17092 memoizedIsInitialized = 1; 17093 return true; 17094 } 17095 17096 public void writeTo(com.google.protobuf.CodedOutputStream output) 17097 throws java.io.IOException { 17098 getSerializedSize(); 17099 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17100 output.writeUInt32(1, snapshotId_); 17101 } 17102 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17103 output.writeUInt64(2, fileSize_); 17104 } 17105 if (((bitField0_ & 0x00000004) == 0x00000004)) { 17106 output.writeBytes(3, name_); 17107 } 17108 if (((bitField0_ & 0x00000008) == 0x00000008)) { 17109 output.writeMessage(4, snapshotCopy_); 17110 } 17111 getUnknownFields().writeTo(output); 17112 } 17113 17114 private int memoizedSerializedSize = -1; 17115 public int getSerializedSize() { 17116 int size = memoizedSerializedSize; 17117 if (size != -1) return size; 17118 17119 size = 0; 17120 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17121 size += com.google.protobuf.CodedOutputStream 17122 .computeUInt32Size(1, snapshotId_); 17123 } 17124 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17125 size += com.google.protobuf.CodedOutputStream 17126 .computeUInt64Size(2, fileSize_); 17127 } 17128 if (((bitField0_ & 0x00000004) == 0x00000004)) { 17129 size += com.google.protobuf.CodedOutputStream 17130 .computeBytesSize(3, name_); 17131 } 17132 if (((bitField0_ & 0x00000008) == 0x00000008)) { 17133 size += com.google.protobuf.CodedOutputStream 17134 .computeMessageSize(4, snapshotCopy_); 17135 } 17136 size += getUnknownFields().getSerializedSize(); 17137 memoizedSerializedSize = size; 17138 return size; 17139 } 17140 17141 private static final long serialVersionUID = 0L; 17142 @java.lang.Override 17143 protected java.lang.Object writeReplace() 17144 throws java.io.ObjectStreamException { 17145 return super.writeReplace(); 17146 } 17147 17148 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17149 com.google.protobuf.ByteString data) 17150 throws com.google.protobuf.InvalidProtocolBufferException { 17151 return PARSER.parseFrom(data); 17152 } 17153 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17154 com.google.protobuf.ByteString data, 17155 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17156 throws com.google.protobuf.InvalidProtocolBufferException { 17157 return PARSER.parseFrom(data, extensionRegistry); 17158 } 17159 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom(byte[] data) 17160 throws com.google.protobuf.InvalidProtocolBufferException { 17161 return PARSER.parseFrom(data); 17162 } 17163 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17164 byte[] data, 17165 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17166 throws com.google.protobuf.InvalidProtocolBufferException { 17167 return PARSER.parseFrom(data, extensionRegistry); 17168 } 17169 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom(java.io.InputStream input) 17170 throws java.io.IOException { 17171 return PARSER.parseFrom(input); 17172 } 17173 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17174 java.io.InputStream input, 17175 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17176 throws java.io.IOException { 17177 return PARSER.parseFrom(input, extensionRegistry); 17178 } 17179 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseDelimitedFrom(java.io.InputStream input) 17180 throws java.io.IOException { 17181 return PARSER.parseDelimitedFrom(input); 17182 } 17183 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseDelimitedFrom( 17184 java.io.InputStream input, 17185 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17186 throws java.io.IOException { 17187 return PARSER.parseDelimitedFrom(input, extensionRegistry); 17188 } 17189 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17190 com.google.protobuf.CodedInputStream input) 17191 throws java.io.IOException { 17192 return PARSER.parseFrom(input); 17193 } 17194 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parseFrom( 17195 com.google.protobuf.CodedInputStream input, 17196 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17197 throws java.io.IOException { 17198 return PARSER.parseFrom(input, extensionRegistry); 17199 } 17200 17201 public static Builder newBuilder() { return Builder.create(); } 17202 public Builder newBuilderForType() { return newBuilder(); } 17203 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff prototype) { 17204 return newBuilder().mergeFrom(prototype); 17205 } 17206 public Builder toBuilder() { return newBuilder(this); } 17207 17208 @java.lang.Override 17209 protected Builder newBuilderForType( 17210 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17211 Builder builder = new Builder(parent); 17212 return builder; 17213 } 17214 /** 17215 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff} 17216 */ 17217 public static final class Builder extends 17218 com.google.protobuf.GeneratedMessage.Builder<Builder> 17219 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiffOrBuilder { 17220 public static final com.google.protobuf.Descriptors.Descriptor 17221 getDescriptor() { 17222 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 17223 } 17224 17225 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17226 internalGetFieldAccessorTable() { 17227 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable 17228 .ensureFieldAccessorsInitialized( 17229 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.Builder.class); 17230 } 17231 17232 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.newBuilder() 17233 private Builder() { 17234 maybeForceBuilderInitialization(); 17235 } 17236 17237 private Builder( 17238 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 17239 super(parent); 17240 maybeForceBuilderInitialization(); 17241 } 17242 private void maybeForceBuilderInitialization() { 17243 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 17244 getSnapshotCopyFieldBuilder(); 17245 } 17246 } 17247 private static Builder create() { 17248 return new Builder(); 17249 } 17250 17251 public Builder clear() { 17252 super.clear(); 17253 snapshotId_ = 0; 17254 bitField0_ = (bitField0_ & ~0x00000001); 17255 fileSize_ = 0L; 17256 bitField0_ = (bitField0_ & ~0x00000002); 17257 name_ = com.google.protobuf.ByteString.EMPTY; 17258 bitField0_ = (bitField0_ & ~0x00000004); 17259 if (snapshotCopyBuilder_ == null) { 17260 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 17261 } else { 17262 snapshotCopyBuilder_.clear(); 17263 } 17264 bitField0_ = (bitField0_ & ~0x00000008); 17265 return this; 17266 } 17267 17268 public Builder clone() { 17269 return create().mergeFrom(buildPartial()); 17270 } 17271 17272 public com.google.protobuf.Descriptors.Descriptor 17273 getDescriptorForType() { 17274 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 17275 } 17276 17277 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff getDefaultInstanceForType() { 17278 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.getDefaultInstance(); 17279 } 17280 17281 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff build() { 17282 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff result = buildPartial(); 17283 if (!result.isInitialized()) { 17284 throw newUninitializedMessageException(result); 17285 } 17286 return result; 17287 } 17288 17289 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff buildPartial() { 17290 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff(this); 17291 int from_bitField0_ = bitField0_; 17292 int to_bitField0_ = 0; 17293 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 17294 to_bitField0_ |= 0x00000001; 17295 } 17296 result.snapshotId_ = snapshotId_; 17297 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 17298 to_bitField0_ |= 0x00000002; 17299 } 17300 result.fileSize_ = fileSize_; 17301 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 17302 to_bitField0_ |= 0x00000004; 17303 } 17304 result.name_ = name_; 17305 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 17306 to_bitField0_ |= 0x00000008; 17307 } 17308 if (snapshotCopyBuilder_ == null) { 17309 result.snapshotCopy_ = snapshotCopy_; 17310 } else { 17311 result.snapshotCopy_ = snapshotCopyBuilder_.build(); 17312 } 17313 result.bitField0_ = to_bitField0_; 17314 onBuilt(); 17315 return result; 17316 } 17317 17318 public Builder mergeFrom(com.google.protobuf.Message other) { 17319 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff) { 17320 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff)other); 17321 } else { 17322 super.mergeFrom(other); 17323 return this; 17324 } 17325 } 17326 17327 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff other) { 17328 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff.getDefaultInstance()) return this; 17329 if (other.hasSnapshotId()) { 17330 setSnapshotId(other.getSnapshotId()); 17331 } 17332 if (other.hasFileSize()) { 17333 setFileSize(other.getFileSize()); 17334 } 17335 if (other.hasName()) { 17336 setName(other.getName()); 17337 } 17338 if (other.hasSnapshotCopy()) { 17339 mergeSnapshotCopy(other.getSnapshotCopy()); 17340 } 17341 this.mergeUnknownFields(other.getUnknownFields()); 17342 return this; 17343 } 17344 17345 public final boolean isInitialized() { 17346 if (hasSnapshotCopy()) { 17347 if (!getSnapshotCopy().isInitialized()) { 17348 17349 return false; 17350 } 17351 } 17352 return true; 17353 } 17354 17355 public Builder mergeFrom( 17356 com.google.protobuf.CodedInputStream input, 17357 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17358 throws java.io.IOException { 17359 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff parsedMessage = null; 17360 try { 17361 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 17362 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17363 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.FileDiff) e.getUnfinishedMessage(); 17364 throw e; 17365 } finally { 17366 if (parsedMessage != null) { 17367 mergeFrom(parsedMessage); 17368 } 17369 } 17370 return this; 17371 } 17372 private int bitField0_; 17373 17374 // optional uint32 snapshotId = 1; 17375 private int snapshotId_ ; 17376 /** 17377 * <code>optional uint32 snapshotId = 1;</code> 17378 */ 17379 public boolean hasSnapshotId() { 17380 return ((bitField0_ & 0x00000001) == 0x00000001); 17381 } 17382 /** 17383 * <code>optional uint32 snapshotId = 1;</code> 17384 */ 17385 public int getSnapshotId() { 17386 return snapshotId_; 17387 } 17388 /** 17389 * <code>optional uint32 snapshotId = 1;</code> 17390 */ 17391 public Builder setSnapshotId(int value) { 17392 bitField0_ |= 0x00000001; 17393 snapshotId_ = value; 17394 onChanged(); 17395 return this; 17396 } 17397 /** 17398 * <code>optional uint32 snapshotId = 1;</code> 17399 */ 17400 public Builder clearSnapshotId() { 17401 bitField0_ = (bitField0_ & ~0x00000001); 17402 snapshotId_ = 0; 17403 onChanged(); 17404 return this; 17405 } 17406 17407 // optional uint64 fileSize = 2; 17408 private long fileSize_ ; 17409 /** 17410 * <code>optional uint64 fileSize = 2;</code> 17411 */ 17412 public boolean hasFileSize() { 17413 return ((bitField0_ & 0x00000002) == 0x00000002); 17414 } 17415 /** 17416 * <code>optional uint64 fileSize = 2;</code> 17417 */ 17418 public long getFileSize() { 17419 return fileSize_; 17420 } 17421 /** 17422 * <code>optional uint64 fileSize = 2;</code> 17423 */ 17424 public Builder setFileSize(long value) { 17425 bitField0_ |= 0x00000002; 17426 fileSize_ = value; 17427 onChanged(); 17428 return this; 17429 } 17430 /** 17431 * <code>optional uint64 fileSize = 2;</code> 17432 */ 17433 public Builder clearFileSize() { 17434 bitField0_ = (bitField0_ & ~0x00000002); 17435 fileSize_ = 0L; 17436 onChanged(); 17437 return this; 17438 } 17439 17440 // optional bytes name = 3; 17441 private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; 17442 /** 17443 * <code>optional bytes name = 3;</code> 17444 */ 17445 public boolean hasName() { 17446 return ((bitField0_ & 0x00000004) == 0x00000004); 17447 } 17448 /** 17449 * <code>optional bytes name = 3;</code> 17450 */ 17451 public com.google.protobuf.ByteString getName() { 17452 return name_; 17453 } 17454 /** 17455 * <code>optional bytes name = 3;</code> 17456 */ 17457 public Builder setName(com.google.protobuf.ByteString value) { 17458 if (value == null) { 17459 throw new NullPointerException(); 17460 } 17461 bitField0_ |= 0x00000004; 17462 name_ = value; 17463 onChanged(); 17464 return this; 17465 } 17466 /** 17467 * <code>optional bytes name = 3;</code> 17468 */ 17469 public Builder clearName() { 17470 bitField0_ = (bitField0_ & ~0x00000004); 17471 name_ = getDefaultInstance().getName(); 17472 onChanged(); 17473 return this; 17474 } 17475 17476 // optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4; 17477 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 17478 private com.google.protobuf.SingleFieldBuilder< 17479 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> snapshotCopyBuilder_; 17480 /** 17481 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17482 */ 17483 public boolean hasSnapshotCopy() { 17484 return ((bitField0_ & 0x00000008) == 0x00000008); 17485 } 17486 /** 17487 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17488 */ 17489 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile getSnapshotCopy() { 17490 if (snapshotCopyBuilder_ == null) { 17491 return snapshotCopy_; 17492 } else { 17493 return snapshotCopyBuilder_.getMessage(); 17494 } 17495 } 17496 /** 17497 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17498 */ 17499 public Builder setSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 17500 if (snapshotCopyBuilder_ == null) { 17501 if (value == null) { 17502 throw new NullPointerException(); 17503 } 17504 snapshotCopy_ = value; 17505 onChanged(); 17506 } else { 17507 snapshotCopyBuilder_.setMessage(value); 17508 } 17509 bitField0_ |= 0x00000008; 17510 return this; 17511 } 17512 /** 17513 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17514 */ 17515 public Builder setSnapshotCopy( 17516 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder builderForValue) { 17517 if (snapshotCopyBuilder_ == null) { 17518 snapshotCopy_ = builderForValue.build(); 17519 onChanged(); 17520 } else { 17521 snapshotCopyBuilder_.setMessage(builderForValue.build()); 17522 } 17523 bitField0_ |= 0x00000008; 17524 return this; 17525 } 17526 /** 17527 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17528 */ 17529 public Builder mergeSnapshotCopy(org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile value) { 17530 if (snapshotCopyBuilder_ == null) { 17531 if (((bitField0_ & 0x00000008) == 0x00000008) && 17532 snapshotCopy_ != org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance()) { 17533 snapshotCopy_ = 17534 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.newBuilder(snapshotCopy_).mergeFrom(value).buildPartial(); 17535 } else { 17536 snapshotCopy_ = value; 17537 } 17538 onChanged(); 17539 } else { 17540 snapshotCopyBuilder_.mergeFrom(value); 17541 } 17542 bitField0_ |= 0x00000008; 17543 return this; 17544 } 17545 /** 17546 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17547 */ 17548 public Builder clearSnapshotCopy() { 17549 if (snapshotCopyBuilder_ == null) { 17550 snapshotCopy_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.getDefaultInstance(); 17551 onChanged(); 17552 } else { 17553 snapshotCopyBuilder_.clear(); 17554 } 17555 bitField0_ = (bitField0_ & ~0x00000008); 17556 return this; 17557 } 17558 /** 17559 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17560 */ 17561 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder getSnapshotCopyBuilder() { 17562 bitField0_ |= 0x00000008; 17563 onChanged(); 17564 return getSnapshotCopyFieldBuilder().getBuilder(); 17565 } 17566 /** 17567 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17568 */ 17569 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder getSnapshotCopyOrBuilder() { 17570 if (snapshotCopyBuilder_ != null) { 17571 return snapshotCopyBuilder_.getMessageOrBuilder(); 17572 } else { 17573 return snapshotCopy_; 17574 } 17575 } 17576 /** 17577 * <code>optional .hadoop.hdfs.fsimage.INodeSection.INodeFile snapshotCopy = 4;</code> 17578 */ 17579 private com.google.protobuf.SingleFieldBuilder< 17580 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder> 17581 getSnapshotCopyFieldBuilder() { 17582 if (snapshotCopyBuilder_ == null) { 17583 snapshotCopyBuilder_ = new com.google.protobuf.SingleFieldBuilder< 17584 org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFile.Builder, org.apache.hadoop.hdfs.server.namenode.FsImageProto.INodeSection.INodeFileOrBuilder>( 17585 snapshotCopy_, 17586 getParentForChildren(), 17587 isClean()); 17588 snapshotCopy_ = null; 17589 } 17590 return snapshotCopyBuilder_; 17591 } 17592 17593 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff) 17594 } 17595 17596 static { 17597 defaultInstance = new FileDiff(true); 17598 defaultInstance.initFields(); 17599 } 17600 17601 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.FileDiff) 17602 } 17603 17604 public interface DiffEntryOrBuilder 17605 extends com.google.protobuf.MessageOrBuilder { 17606 17607 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 17608 /** 17609 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 17610 */ 17611 boolean hasType(); 17612 /** 17613 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 17614 */ 17615 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType(); 17616 17617 // optional uint64 inodeId = 2; 17618 /** 17619 * <code>optional uint64 inodeId = 2;</code> 17620 */ 17621 boolean hasInodeId(); 17622 /** 17623 * <code>optional uint64 inodeId = 2;</code> 17624 */ 17625 long getInodeId(); 17626 17627 // optional uint32 numOfDiff = 3; 17628 /** 17629 * <code>optional uint32 numOfDiff = 3;</code> 17630 */ 17631 boolean hasNumOfDiff(); 17632 /** 17633 * <code>optional uint32 numOfDiff = 3;</code> 17634 */ 17635 int getNumOfDiff(); 17636 } 17637 /** 17638 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry} 17639 */ 17640 public static final class DiffEntry extends 17641 com.google.protobuf.GeneratedMessage 17642 implements DiffEntryOrBuilder { 17643 // Use DiffEntry.newBuilder() to construct. 17644 private DiffEntry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 17645 super(builder); 17646 this.unknownFields = builder.getUnknownFields(); 17647 } 17648 private DiffEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 17649 17650 private static final DiffEntry defaultInstance; 17651 public static DiffEntry getDefaultInstance() { 17652 return defaultInstance; 17653 } 17654 17655 public DiffEntry getDefaultInstanceForType() { 17656 return defaultInstance; 17657 } 17658 17659 private final com.google.protobuf.UnknownFieldSet unknownFields; 17660 @java.lang.Override 17661 public final com.google.protobuf.UnknownFieldSet 17662 getUnknownFields() { 17663 return this.unknownFields; 17664 } 17665 private DiffEntry( 17666 com.google.protobuf.CodedInputStream input, 17667 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17668 throws com.google.protobuf.InvalidProtocolBufferException { 17669 initFields(); 17670 int mutable_bitField0_ = 0; 17671 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 17672 com.google.protobuf.UnknownFieldSet.newBuilder(); 17673 try { 17674 boolean done = false; 17675 while (!done) { 17676 int tag = input.readTag(); 17677 switch (tag) { 17678 case 0: 17679 done = true; 17680 break; 17681 default: { 17682 if (!parseUnknownField(input, unknownFields, 17683 extensionRegistry, tag)) { 17684 done = true; 17685 } 17686 break; 17687 } 17688 case 8: { 17689 int rawValue = input.readEnum(); 17690 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type value = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.valueOf(rawValue); 17691 if (value == null) { 17692 unknownFields.mergeVarintField(1, rawValue); 17693 } else { 17694 bitField0_ |= 0x00000001; 17695 type_ = value; 17696 } 17697 break; 17698 } 17699 case 16: { 17700 bitField0_ |= 0x00000002; 17701 inodeId_ = input.readUInt64(); 17702 break; 17703 } 17704 case 24: { 17705 bitField0_ |= 0x00000004; 17706 numOfDiff_ = input.readUInt32(); 17707 break; 17708 } 17709 } 17710 } 17711 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 17712 throw e.setUnfinishedMessage(this); 17713 } catch (java.io.IOException e) { 17714 throw new com.google.protobuf.InvalidProtocolBufferException( 17715 e.getMessage()).setUnfinishedMessage(this); 17716 } finally { 17717 this.unknownFields = unknownFields.build(); 17718 makeExtensionsImmutable(); 17719 } 17720 } 17721 public static final com.google.protobuf.Descriptors.Descriptor 17722 getDescriptor() { 17723 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 17724 } 17725 17726 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 17727 internalGetFieldAccessorTable() { 17728 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable 17729 .ensureFieldAccessorsInitialized( 17730 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Builder.class); 17731 } 17732 17733 public static com.google.protobuf.Parser<DiffEntry> PARSER = 17734 new com.google.protobuf.AbstractParser<DiffEntry>() { 17735 public DiffEntry parsePartialFrom( 17736 com.google.protobuf.CodedInputStream input, 17737 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17738 throws com.google.protobuf.InvalidProtocolBufferException { 17739 return new DiffEntry(input, extensionRegistry); 17740 } 17741 }; 17742 17743 @java.lang.Override 17744 public com.google.protobuf.Parser<DiffEntry> getParserForType() { 17745 return PARSER; 17746 } 17747 17748 /** 17749 * Protobuf enum {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type} 17750 */ 17751 public enum Type 17752 implements com.google.protobuf.ProtocolMessageEnum { 17753 /** 17754 * <code>FILEDIFF = 1;</code> 17755 */ 17756 FILEDIFF(0, 1), 17757 /** 17758 * <code>DIRECTORYDIFF = 2;</code> 17759 */ 17760 DIRECTORYDIFF(1, 2), 17761 ; 17762 17763 /** 17764 * <code>FILEDIFF = 1;</code> 17765 */ 17766 public static final int FILEDIFF_VALUE = 1; 17767 /** 17768 * <code>DIRECTORYDIFF = 2;</code> 17769 */ 17770 public static final int DIRECTORYDIFF_VALUE = 2; 17771 17772 17773 public final int getNumber() { return value; } 17774 17775 public static Type valueOf(int value) { 17776 switch (value) { 17777 case 1: return FILEDIFF; 17778 case 2: return DIRECTORYDIFF; 17779 default: return null; 17780 } 17781 } 17782 17783 public static com.google.protobuf.Internal.EnumLiteMap<Type> 17784 internalGetValueMap() { 17785 return internalValueMap; 17786 } 17787 private static com.google.protobuf.Internal.EnumLiteMap<Type> 17788 internalValueMap = 17789 new com.google.protobuf.Internal.EnumLiteMap<Type>() { 17790 public Type findValueByNumber(int number) { 17791 return Type.valueOf(number); 17792 } 17793 }; 17794 17795 public final com.google.protobuf.Descriptors.EnumValueDescriptor 17796 getValueDescriptor() { 17797 return getDescriptor().getValues().get(index); 17798 } 17799 public final com.google.protobuf.Descriptors.EnumDescriptor 17800 getDescriptorForType() { 17801 return getDescriptor(); 17802 } 17803 public static final com.google.protobuf.Descriptors.EnumDescriptor 17804 getDescriptor() { 17805 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDescriptor().getEnumTypes().get(0); 17806 } 17807 17808 private static final Type[] VALUES = values(); 17809 17810 public static Type valueOf( 17811 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 17812 if (desc.getType() != getDescriptor()) { 17813 throw new java.lang.IllegalArgumentException( 17814 "EnumValueDescriptor is not for this type."); 17815 } 17816 return VALUES[desc.getIndex()]; 17817 } 17818 17819 private final int index; 17820 private final int value; 17821 17822 private Type(int index, int value) { 17823 this.index = index; 17824 this.value = value; 17825 } 17826 17827 // @@protoc_insertion_point(enum_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type) 17828 } 17829 17830 private int bitField0_; 17831 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 17832 public static final int TYPE_FIELD_NUMBER = 1; 17833 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type type_; 17834 /** 17835 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 17836 */ 17837 public boolean hasType() { 17838 return ((bitField0_ & 0x00000001) == 0x00000001); 17839 } 17840 /** 17841 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 17842 */ 17843 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType() { 17844 return type_; 17845 } 17846 17847 // optional uint64 inodeId = 2; 17848 public static final int INODEID_FIELD_NUMBER = 2; 17849 private long inodeId_; 17850 /** 17851 * <code>optional uint64 inodeId = 2;</code> 17852 */ 17853 public boolean hasInodeId() { 17854 return ((bitField0_ & 0x00000002) == 0x00000002); 17855 } 17856 /** 17857 * <code>optional uint64 inodeId = 2;</code> 17858 */ 17859 public long getInodeId() { 17860 return inodeId_; 17861 } 17862 17863 // optional uint32 numOfDiff = 3; 17864 public static final int NUMOFDIFF_FIELD_NUMBER = 3; 17865 private int numOfDiff_; 17866 /** 17867 * <code>optional uint32 numOfDiff = 3;</code> 17868 */ 17869 public boolean hasNumOfDiff() { 17870 return ((bitField0_ & 0x00000004) == 0x00000004); 17871 } 17872 /** 17873 * <code>optional uint32 numOfDiff = 3;</code> 17874 */ 17875 public int getNumOfDiff() { 17876 return numOfDiff_; 17877 } 17878 17879 private void initFields() { 17880 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 17881 inodeId_ = 0L; 17882 numOfDiff_ = 0; 17883 } 17884 private byte memoizedIsInitialized = -1; 17885 public final boolean isInitialized() { 17886 byte isInitialized = memoizedIsInitialized; 17887 if (isInitialized != -1) return isInitialized == 1; 17888 17889 if (!hasType()) { 17890 memoizedIsInitialized = 0; 17891 return false; 17892 } 17893 memoizedIsInitialized = 1; 17894 return true; 17895 } 17896 17897 public void writeTo(com.google.protobuf.CodedOutputStream output) 17898 throws java.io.IOException { 17899 getSerializedSize(); 17900 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17901 output.writeEnum(1, type_.getNumber()); 17902 } 17903 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17904 output.writeUInt64(2, inodeId_); 17905 } 17906 if (((bitField0_ & 0x00000004) == 0x00000004)) { 17907 output.writeUInt32(3, numOfDiff_); 17908 } 17909 getUnknownFields().writeTo(output); 17910 } 17911 17912 private int memoizedSerializedSize = -1; 17913 public int getSerializedSize() { 17914 int size = memoizedSerializedSize; 17915 if (size != -1) return size; 17916 17917 size = 0; 17918 if (((bitField0_ & 0x00000001) == 0x00000001)) { 17919 size += com.google.protobuf.CodedOutputStream 17920 .computeEnumSize(1, type_.getNumber()); 17921 } 17922 if (((bitField0_ & 0x00000002) == 0x00000002)) { 17923 size += com.google.protobuf.CodedOutputStream 17924 .computeUInt64Size(2, inodeId_); 17925 } 17926 if (((bitField0_ & 0x00000004) == 0x00000004)) { 17927 size += com.google.protobuf.CodedOutputStream 17928 .computeUInt32Size(3, numOfDiff_); 17929 } 17930 size += getUnknownFields().getSerializedSize(); 17931 memoizedSerializedSize = size; 17932 return size; 17933 } 17934 17935 private static final long serialVersionUID = 0L; 17936 @java.lang.Override 17937 protected java.lang.Object writeReplace() 17938 throws java.io.ObjectStreamException { 17939 return super.writeReplace(); 17940 } 17941 17942 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17943 com.google.protobuf.ByteString data) 17944 throws com.google.protobuf.InvalidProtocolBufferException { 17945 return PARSER.parseFrom(data); 17946 } 17947 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17948 com.google.protobuf.ByteString data, 17949 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17950 throws com.google.protobuf.InvalidProtocolBufferException { 17951 return PARSER.parseFrom(data, extensionRegistry); 17952 } 17953 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom(byte[] data) 17954 throws com.google.protobuf.InvalidProtocolBufferException { 17955 return PARSER.parseFrom(data); 17956 } 17957 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17958 byte[] data, 17959 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17960 throws com.google.protobuf.InvalidProtocolBufferException { 17961 return PARSER.parseFrom(data, extensionRegistry); 17962 } 17963 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom(java.io.InputStream input) 17964 throws java.io.IOException { 17965 return PARSER.parseFrom(input); 17966 } 17967 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17968 java.io.InputStream input, 17969 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17970 throws java.io.IOException { 17971 return PARSER.parseFrom(input, extensionRegistry); 17972 } 17973 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseDelimitedFrom(java.io.InputStream input) 17974 throws java.io.IOException { 17975 return PARSER.parseDelimitedFrom(input); 17976 } 17977 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseDelimitedFrom( 17978 java.io.InputStream input, 17979 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17980 throws java.io.IOException { 17981 return PARSER.parseDelimitedFrom(input, extensionRegistry); 17982 } 17983 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17984 com.google.protobuf.CodedInputStream input) 17985 throws java.io.IOException { 17986 return PARSER.parseFrom(input); 17987 } 17988 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parseFrom( 17989 com.google.protobuf.CodedInputStream input, 17990 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 17991 throws java.io.IOException { 17992 return PARSER.parseFrom(input, extensionRegistry); 17993 } 17994 17995 public static Builder newBuilder() { return Builder.create(); } 17996 public Builder newBuilderForType() { return newBuilder(); } 17997 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry prototype) { 17998 return newBuilder().mergeFrom(prototype); 17999 } 18000 public Builder toBuilder() { return newBuilder(this); } 18001 18002 @java.lang.Override 18003 protected Builder newBuilderForType( 18004 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18005 Builder builder = new Builder(parent); 18006 return builder; 18007 } 18008 /** 18009 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry} 18010 */ 18011 public static final class Builder extends 18012 com.google.protobuf.GeneratedMessage.Builder<Builder> 18013 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntryOrBuilder { 18014 public static final com.google.protobuf.Descriptors.Descriptor 18015 getDescriptor() { 18016 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 18017 } 18018 18019 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18020 internalGetFieldAccessorTable() { 18021 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable 18022 .ensureFieldAccessorsInitialized( 18023 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Builder.class); 18024 } 18025 18026 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.newBuilder() 18027 private Builder() { 18028 maybeForceBuilderInitialization(); 18029 } 18030 18031 private Builder( 18032 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18033 super(parent); 18034 maybeForceBuilderInitialization(); 18035 } 18036 private void maybeForceBuilderInitialization() { 18037 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 18038 } 18039 } 18040 private static Builder create() { 18041 return new Builder(); 18042 } 18043 18044 public Builder clear() { 18045 super.clear(); 18046 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 18047 bitField0_ = (bitField0_ & ~0x00000001); 18048 inodeId_ = 0L; 18049 bitField0_ = (bitField0_ & ~0x00000002); 18050 numOfDiff_ = 0; 18051 bitField0_ = (bitField0_ & ~0x00000004); 18052 return this; 18053 } 18054 18055 public Builder clone() { 18056 return create().mergeFrom(buildPartial()); 18057 } 18058 18059 public com.google.protobuf.Descriptors.Descriptor 18060 getDescriptorForType() { 18061 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 18062 } 18063 18064 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry getDefaultInstanceForType() { 18065 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDefaultInstance(); 18066 } 18067 18068 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry build() { 18069 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry result = buildPartial(); 18070 if (!result.isInitialized()) { 18071 throw newUninitializedMessageException(result); 18072 } 18073 return result; 18074 } 18075 18076 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry buildPartial() { 18077 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry(this); 18078 int from_bitField0_ = bitField0_; 18079 int to_bitField0_ = 0; 18080 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 18081 to_bitField0_ |= 0x00000001; 18082 } 18083 result.type_ = type_; 18084 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 18085 to_bitField0_ |= 0x00000002; 18086 } 18087 result.inodeId_ = inodeId_; 18088 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 18089 to_bitField0_ |= 0x00000004; 18090 } 18091 result.numOfDiff_ = numOfDiff_; 18092 result.bitField0_ = to_bitField0_; 18093 onBuilt(); 18094 return result; 18095 } 18096 18097 public Builder mergeFrom(com.google.protobuf.Message other) { 18098 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry) { 18099 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry)other); 18100 } else { 18101 super.mergeFrom(other); 18102 return this; 18103 } 18104 } 18105 18106 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry other) { 18107 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.getDefaultInstance()) return this; 18108 if (other.hasType()) { 18109 setType(other.getType()); 18110 } 18111 if (other.hasInodeId()) { 18112 setInodeId(other.getInodeId()); 18113 } 18114 if (other.hasNumOfDiff()) { 18115 setNumOfDiff(other.getNumOfDiff()); 18116 } 18117 this.mergeUnknownFields(other.getUnknownFields()); 18118 return this; 18119 } 18120 18121 public final boolean isInitialized() { 18122 if (!hasType()) { 18123 18124 return false; 18125 } 18126 return true; 18127 } 18128 18129 public Builder mergeFrom( 18130 com.google.protobuf.CodedInputStream input, 18131 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18132 throws java.io.IOException { 18133 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry parsedMessage = null; 18134 try { 18135 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 18136 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18137 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry) e.getUnfinishedMessage(); 18138 throw e; 18139 } finally { 18140 if (parsedMessage != null) { 18141 mergeFrom(parsedMessage); 18142 } 18143 } 18144 return this; 18145 } 18146 private int bitField0_; 18147 18148 // required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1; 18149 private org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 18150 /** 18151 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 18152 */ 18153 public boolean hasType() { 18154 return ((bitField0_ & 0x00000001) == 0x00000001); 18155 } 18156 /** 18157 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 18158 */ 18159 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type getType() { 18160 return type_; 18161 } 18162 /** 18163 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 18164 */ 18165 public Builder setType(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type value) { 18166 if (value == null) { 18167 throw new NullPointerException(); 18168 } 18169 bitField0_ |= 0x00000001; 18170 type_ = value; 18171 onChanged(); 18172 return this; 18173 } 18174 /** 18175 * <code>required .hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry.Type type = 1;</code> 18176 */ 18177 public Builder clearType() { 18178 bitField0_ = (bitField0_ & ~0x00000001); 18179 type_ = org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.DiffEntry.Type.FILEDIFF; 18180 onChanged(); 18181 return this; 18182 } 18183 18184 // optional uint64 inodeId = 2; 18185 private long inodeId_ ; 18186 /** 18187 * <code>optional uint64 inodeId = 2;</code> 18188 */ 18189 public boolean hasInodeId() { 18190 return ((bitField0_ & 0x00000002) == 0x00000002); 18191 } 18192 /** 18193 * <code>optional uint64 inodeId = 2;</code> 18194 */ 18195 public long getInodeId() { 18196 return inodeId_; 18197 } 18198 /** 18199 * <code>optional uint64 inodeId = 2;</code> 18200 */ 18201 public Builder setInodeId(long value) { 18202 bitField0_ |= 0x00000002; 18203 inodeId_ = value; 18204 onChanged(); 18205 return this; 18206 } 18207 /** 18208 * <code>optional uint64 inodeId = 2;</code> 18209 */ 18210 public Builder clearInodeId() { 18211 bitField0_ = (bitField0_ & ~0x00000002); 18212 inodeId_ = 0L; 18213 onChanged(); 18214 return this; 18215 } 18216 18217 // optional uint32 numOfDiff = 3; 18218 private int numOfDiff_ ; 18219 /** 18220 * <code>optional uint32 numOfDiff = 3;</code> 18221 */ 18222 public boolean hasNumOfDiff() { 18223 return ((bitField0_ & 0x00000004) == 0x00000004); 18224 } 18225 /** 18226 * <code>optional uint32 numOfDiff = 3;</code> 18227 */ 18228 public int getNumOfDiff() { 18229 return numOfDiff_; 18230 } 18231 /** 18232 * <code>optional uint32 numOfDiff = 3;</code> 18233 */ 18234 public Builder setNumOfDiff(int value) { 18235 bitField0_ |= 0x00000004; 18236 numOfDiff_ = value; 18237 onChanged(); 18238 return this; 18239 } 18240 /** 18241 * <code>optional uint32 numOfDiff = 3;</code> 18242 */ 18243 public Builder clearNumOfDiff() { 18244 bitField0_ = (bitField0_ & ~0x00000004); 18245 numOfDiff_ = 0; 18246 onChanged(); 18247 return this; 18248 } 18249 18250 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry) 18251 } 18252 18253 static { 18254 defaultInstance = new DiffEntry(true); 18255 defaultInstance.initFields(); 18256 } 18257 18258 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection.DiffEntry) 18259 } 18260 18261 private void initFields() { 18262 } 18263 private byte memoizedIsInitialized = -1; 18264 public final boolean isInitialized() { 18265 byte isInitialized = memoizedIsInitialized; 18266 if (isInitialized != -1) return isInitialized == 1; 18267 18268 memoizedIsInitialized = 1; 18269 return true; 18270 } 18271 18272 public void writeTo(com.google.protobuf.CodedOutputStream output) 18273 throws java.io.IOException { 18274 getSerializedSize(); 18275 getUnknownFields().writeTo(output); 18276 } 18277 18278 private int memoizedSerializedSize = -1; 18279 public int getSerializedSize() { 18280 int size = memoizedSerializedSize; 18281 if (size != -1) return size; 18282 18283 size = 0; 18284 size += getUnknownFields().getSerializedSize(); 18285 memoizedSerializedSize = size; 18286 return size; 18287 } 18288 18289 private static final long serialVersionUID = 0L; 18290 @java.lang.Override 18291 protected java.lang.Object writeReplace() 18292 throws java.io.ObjectStreamException { 18293 return super.writeReplace(); 18294 } 18295 18296 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18297 com.google.protobuf.ByteString data) 18298 throws com.google.protobuf.InvalidProtocolBufferException { 18299 return PARSER.parseFrom(data); 18300 } 18301 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18302 com.google.protobuf.ByteString data, 18303 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18304 throws com.google.protobuf.InvalidProtocolBufferException { 18305 return PARSER.parseFrom(data, extensionRegistry); 18306 } 18307 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom(byte[] data) 18308 throws com.google.protobuf.InvalidProtocolBufferException { 18309 return PARSER.parseFrom(data); 18310 } 18311 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18312 byte[] data, 18313 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18314 throws com.google.protobuf.InvalidProtocolBufferException { 18315 return PARSER.parseFrom(data, extensionRegistry); 18316 } 18317 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom(java.io.InputStream input) 18318 throws java.io.IOException { 18319 return PARSER.parseFrom(input); 18320 } 18321 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18322 java.io.InputStream input, 18323 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18324 throws java.io.IOException { 18325 return PARSER.parseFrom(input, extensionRegistry); 18326 } 18327 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseDelimitedFrom(java.io.InputStream input) 18328 throws java.io.IOException { 18329 return PARSER.parseDelimitedFrom(input); 18330 } 18331 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseDelimitedFrom( 18332 java.io.InputStream input, 18333 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18334 throws java.io.IOException { 18335 return PARSER.parseDelimitedFrom(input, extensionRegistry); 18336 } 18337 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18338 com.google.protobuf.CodedInputStream input) 18339 throws java.io.IOException { 18340 return PARSER.parseFrom(input); 18341 } 18342 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parseFrom( 18343 com.google.protobuf.CodedInputStream input, 18344 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18345 throws java.io.IOException { 18346 return PARSER.parseFrom(input, extensionRegistry); 18347 } 18348 18349 public static Builder newBuilder() { return Builder.create(); } 18350 public Builder newBuilderForType() { return newBuilder(); } 18351 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection prototype) { 18352 return newBuilder().mergeFrom(prototype); 18353 } 18354 public Builder toBuilder() { return newBuilder(this); } 18355 18356 @java.lang.Override 18357 protected Builder newBuilderForType( 18358 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18359 Builder builder = new Builder(parent); 18360 return builder; 18361 } 18362 /** 18363 * Protobuf type {@code hadoop.hdfs.fsimage.SnapshotDiffSection} 18364 * 18365 * <pre> 18366 ** 18367 * This section records information about snapshot diffs 18368 * NAME: SNAPSHOT_DIFF 18369 * </pre> 18370 */ 18371 public static final class Builder extends 18372 com.google.protobuf.GeneratedMessage.Builder<Builder> 18373 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSectionOrBuilder { 18374 public static final com.google.protobuf.Descriptors.Descriptor 18375 getDescriptor() { 18376 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 18377 } 18378 18379 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18380 internalGetFieldAccessorTable() { 18381 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable 18382 .ensureFieldAccessorsInitialized( 18383 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.Builder.class); 18384 } 18385 18386 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.newBuilder() 18387 private Builder() { 18388 maybeForceBuilderInitialization(); 18389 } 18390 18391 private Builder( 18392 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18393 super(parent); 18394 maybeForceBuilderInitialization(); 18395 } 18396 private void maybeForceBuilderInitialization() { 18397 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 18398 } 18399 } 18400 private static Builder create() { 18401 return new Builder(); 18402 } 18403 18404 public Builder clear() { 18405 super.clear(); 18406 return this; 18407 } 18408 18409 public Builder clone() { 18410 return create().mergeFrom(buildPartial()); 18411 } 18412 18413 public com.google.protobuf.Descriptors.Descriptor 18414 getDescriptorForType() { 18415 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 18416 } 18417 18418 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection getDefaultInstanceForType() { 18419 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.getDefaultInstance(); 18420 } 18421 18422 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection build() { 18423 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection result = buildPartial(); 18424 if (!result.isInitialized()) { 18425 throw newUninitializedMessageException(result); 18426 } 18427 return result; 18428 } 18429 18430 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection buildPartial() { 18431 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection(this); 18432 onBuilt(); 18433 return result; 18434 } 18435 18436 public Builder mergeFrom(com.google.protobuf.Message other) { 18437 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection) { 18438 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection)other); 18439 } else { 18440 super.mergeFrom(other); 18441 return this; 18442 } 18443 } 18444 18445 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection other) { 18446 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection.getDefaultInstance()) return this; 18447 this.mergeUnknownFields(other.getUnknownFields()); 18448 return this; 18449 } 18450 18451 public final boolean isInitialized() { 18452 return true; 18453 } 18454 18455 public Builder mergeFrom( 18456 com.google.protobuf.CodedInputStream input, 18457 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18458 throws java.io.IOException { 18459 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection parsedMessage = null; 18460 try { 18461 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 18462 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18463 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SnapshotDiffSection) e.getUnfinishedMessage(); 18464 throw e; 18465 } finally { 18466 if (parsedMessage != null) { 18467 mergeFrom(parsedMessage); 18468 } 18469 } 18470 return this; 18471 } 18472 18473 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SnapshotDiffSection) 18474 } 18475 18476 static { 18477 defaultInstance = new SnapshotDiffSection(true); 18478 defaultInstance.initFields(); 18479 } 18480 18481 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SnapshotDiffSection) 18482 } 18483 18484 public interface StringTableSectionOrBuilder 18485 extends com.google.protobuf.MessageOrBuilder { 18486 18487 // optional uint32 numEntry = 1; 18488 /** 18489 * <code>optional uint32 numEntry = 1;</code> 18490 * 18491 * <pre> 18492 * repeated Entry 18493 * </pre> 18494 */ 18495 boolean hasNumEntry(); 18496 /** 18497 * <code>optional uint32 numEntry = 1;</code> 18498 * 18499 * <pre> 18500 * repeated Entry 18501 * </pre> 18502 */ 18503 int getNumEntry(); 18504 } 18505 /** 18506 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection} 18507 * 18508 * <pre> 18509 ** 18510 * This section maps string to id 18511 * NAME: STRING_TABLE 18512 * </pre> 18513 */ 18514 public static final class StringTableSection extends 18515 com.google.protobuf.GeneratedMessage 18516 implements StringTableSectionOrBuilder { 18517 // Use StringTableSection.newBuilder() to construct. 18518 private StringTableSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 18519 super(builder); 18520 this.unknownFields = builder.getUnknownFields(); 18521 } 18522 private StringTableSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 18523 18524 private static final StringTableSection defaultInstance; 18525 public static StringTableSection getDefaultInstance() { 18526 return defaultInstance; 18527 } 18528 18529 public StringTableSection getDefaultInstanceForType() { 18530 return defaultInstance; 18531 } 18532 18533 private final com.google.protobuf.UnknownFieldSet unknownFields; 18534 @java.lang.Override 18535 public final com.google.protobuf.UnknownFieldSet 18536 getUnknownFields() { 18537 return this.unknownFields; 18538 } 18539 private StringTableSection( 18540 com.google.protobuf.CodedInputStream input, 18541 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18542 throws com.google.protobuf.InvalidProtocolBufferException { 18543 initFields(); 18544 int mutable_bitField0_ = 0; 18545 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 18546 com.google.protobuf.UnknownFieldSet.newBuilder(); 18547 try { 18548 boolean done = false; 18549 while (!done) { 18550 int tag = input.readTag(); 18551 switch (tag) { 18552 case 0: 18553 done = true; 18554 break; 18555 default: { 18556 if (!parseUnknownField(input, unknownFields, 18557 extensionRegistry, tag)) { 18558 done = true; 18559 } 18560 break; 18561 } 18562 case 8: { 18563 bitField0_ |= 0x00000001; 18564 numEntry_ = input.readUInt32(); 18565 break; 18566 } 18567 } 18568 } 18569 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18570 throw e.setUnfinishedMessage(this); 18571 } catch (java.io.IOException e) { 18572 throw new com.google.protobuf.InvalidProtocolBufferException( 18573 e.getMessage()).setUnfinishedMessage(this); 18574 } finally { 18575 this.unknownFields = unknownFields.build(); 18576 makeExtensionsImmutable(); 18577 } 18578 } 18579 public static final com.google.protobuf.Descriptors.Descriptor 18580 getDescriptor() { 18581 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 18582 } 18583 18584 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18585 internalGetFieldAccessorTable() { 18586 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable 18587 .ensureFieldAccessorsInitialized( 18588 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Builder.class); 18589 } 18590 18591 public static com.google.protobuf.Parser<StringTableSection> PARSER = 18592 new com.google.protobuf.AbstractParser<StringTableSection>() { 18593 public StringTableSection parsePartialFrom( 18594 com.google.protobuf.CodedInputStream input, 18595 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18596 throws com.google.protobuf.InvalidProtocolBufferException { 18597 return new StringTableSection(input, extensionRegistry); 18598 } 18599 }; 18600 18601 @java.lang.Override 18602 public com.google.protobuf.Parser<StringTableSection> getParserForType() { 18603 return PARSER; 18604 } 18605 18606 public interface EntryOrBuilder 18607 extends com.google.protobuf.MessageOrBuilder { 18608 18609 // optional uint32 id = 1; 18610 /** 18611 * <code>optional uint32 id = 1;</code> 18612 */ 18613 boolean hasId(); 18614 /** 18615 * <code>optional uint32 id = 1;</code> 18616 */ 18617 int getId(); 18618 18619 // optional string str = 2; 18620 /** 18621 * <code>optional string str = 2;</code> 18622 */ 18623 boolean hasStr(); 18624 /** 18625 * <code>optional string str = 2;</code> 18626 */ 18627 java.lang.String getStr(); 18628 /** 18629 * <code>optional string str = 2;</code> 18630 */ 18631 com.google.protobuf.ByteString 18632 getStrBytes(); 18633 } 18634 /** 18635 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection.Entry} 18636 */ 18637 public static final class Entry extends 18638 com.google.protobuf.GeneratedMessage 18639 implements EntryOrBuilder { 18640 // Use Entry.newBuilder() to construct. 18641 private Entry(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 18642 super(builder); 18643 this.unknownFields = builder.getUnknownFields(); 18644 } 18645 private Entry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 18646 18647 private static final Entry defaultInstance; 18648 public static Entry getDefaultInstance() { 18649 return defaultInstance; 18650 } 18651 18652 public Entry getDefaultInstanceForType() { 18653 return defaultInstance; 18654 } 18655 18656 private final com.google.protobuf.UnknownFieldSet unknownFields; 18657 @java.lang.Override 18658 public final com.google.protobuf.UnknownFieldSet 18659 getUnknownFields() { 18660 return this.unknownFields; 18661 } 18662 private Entry( 18663 com.google.protobuf.CodedInputStream input, 18664 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18665 throws com.google.protobuf.InvalidProtocolBufferException { 18666 initFields(); 18667 int mutable_bitField0_ = 0; 18668 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 18669 com.google.protobuf.UnknownFieldSet.newBuilder(); 18670 try { 18671 boolean done = false; 18672 while (!done) { 18673 int tag = input.readTag(); 18674 switch (tag) { 18675 case 0: 18676 done = true; 18677 break; 18678 default: { 18679 if (!parseUnknownField(input, unknownFields, 18680 extensionRegistry, tag)) { 18681 done = true; 18682 } 18683 break; 18684 } 18685 case 8: { 18686 bitField0_ |= 0x00000001; 18687 id_ = input.readUInt32(); 18688 break; 18689 } 18690 case 18: { 18691 bitField0_ |= 0x00000002; 18692 str_ = input.readBytes(); 18693 break; 18694 } 18695 } 18696 } 18697 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 18698 throw e.setUnfinishedMessage(this); 18699 } catch (java.io.IOException e) { 18700 throw new com.google.protobuf.InvalidProtocolBufferException( 18701 e.getMessage()).setUnfinishedMessage(this); 18702 } finally { 18703 this.unknownFields = unknownFields.build(); 18704 makeExtensionsImmutable(); 18705 } 18706 } 18707 public static final com.google.protobuf.Descriptors.Descriptor 18708 getDescriptor() { 18709 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 18710 } 18711 18712 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18713 internalGetFieldAccessorTable() { 18714 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable 18715 .ensureFieldAccessorsInitialized( 18716 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.Builder.class); 18717 } 18718 18719 public static com.google.protobuf.Parser<Entry> PARSER = 18720 new com.google.protobuf.AbstractParser<Entry>() { 18721 public Entry parsePartialFrom( 18722 com.google.protobuf.CodedInputStream input, 18723 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18724 throws com.google.protobuf.InvalidProtocolBufferException { 18725 return new Entry(input, extensionRegistry); 18726 } 18727 }; 18728 18729 @java.lang.Override 18730 public com.google.protobuf.Parser<Entry> getParserForType() { 18731 return PARSER; 18732 } 18733 18734 private int bitField0_; 18735 // optional uint32 id = 1; 18736 public static final int ID_FIELD_NUMBER = 1; 18737 private int id_; 18738 /** 18739 * <code>optional uint32 id = 1;</code> 18740 */ 18741 public boolean hasId() { 18742 return ((bitField0_ & 0x00000001) == 0x00000001); 18743 } 18744 /** 18745 * <code>optional uint32 id = 1;</code> 18746 */ 18747 public int getId() { 18748 return id_; 18749 } 18750 18751 // optional string str = 2; 18752 public static final int STR_FIELD_NUMBER = 2; 18753 private java.lang.Object str_; 18754 /** 18755 * <code>optional string str = 2;</code> 18756 */ 18757 public boolean hasStr() { 18758 return ((bitField0_ & 0x00000002) == 0x00000002); 18759 } 18760 /** 18761 * <code>optional string str = 2;</code> 18762 */ 18763 public java.lang.String getStr() { 18764 java.lang.Object ref = str_; 18765 if (ref instanceof java.lang.String) { 18766 return (java.lang.String) ref; 18767 } else { 18768 com.google.protobuf.ByteString bs = 18769 (com.google.protobuf.ByteString) ref; 18770 java.lang.String s = bs.toStringUtf8(); 18771 if (bs.isValidUtf8()) { 18772 str_ = s; 18773 } 18774 return s; 18775 } 18776 } 18777 /** 18778 * <code>optional string str = 2;</code> 18779 */ 18780 public com.google.protobuf.ByteString 18781 getStrBytes() { 18782 java.lang.Object ref = str_; 18783 if (ref instanceof java.lang.String) { 18784 com.google.protobuf.ByteString b = 18785 com.google.protobuf.ByteString.copyFromUtf8( 18786 (java.lang.String) ref); 18787 str_ = b; 18788 return b; 18789 } else { 18790 return (com.google.protobuf.ByteString) ref; 18791 } 18792 } 18793 18794 private void initFields() { 18795 id_ = 0; 18796 str_ = ""; 18797 } 18798 private byte memoizedIsInitialized = -1; 18799 public final boolean isInitialized() { 18800 byte isInitialized = memoizedIsInitialized; 18801 if (isInitialized != -1) return isInitialized == 1; 18802 18803 memoizedIsInitialized = 1; 18804 return true; 18805 } 18806 18807 public void writeTo(com.google.protobuf.CodedOutputStream output) 18808 throws java.io.IOException { 18809 getSerializedSize(); 18810 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18811 output.writeUInt32(1, id_); 18812 } 18813 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18814 output.writeBytes(2, getStrBytes()); 18815 } 18816 getUnknownFields().writeTo(output); 18817 } 18818 18819 private int memoizedSerializedSize = -1; 18820 public int getSerializedSize() { 18821 int size = memoizedSerializedSize; 18822 if (size != -1) return size; 18823 18824 size = 0; 18825 if (((bitField0_ & 0x00000001) == 0x00000001)) { 18826 size += com.google.protobuf.CodedOutputStream 18827 .computeUInt32Size(1, id_); 18828 } 18829 if (((bitField0_ & 0x00000002) == 0x00000002)) { 18830 size += com.google.protobuf.CodedOutputStream 18831 .computeBytesSize(2, getStrBytes()); 18832 } 18833 size += getUnknownFields().getSerializedSize(); 18834 memoizedSerializedSize = size; 18835 return size; 18836 } 18837 18838 private static final long serialVersionUID = 0L; 18839 @java.lang.Override 18840 protected java.lang.Object writeReplace() 18841 throws java.io.ObjectStreamException { 18842 return super.writeReplace(); 18843 } 18844 18845 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18846 com.google.protobuf.ByteString data) 18847 throws com.google.protobuf.InvalidProtocolBufferException { 18848 return PARSER.parseFrom(data); 18849 } 18850 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18851 com.google.protobuf.ByteString data, 18852 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18853 throws com.google.protobuf.InvalidProtocolBufferException { 18854 return PARSER.parseFrom(data, extensionRegistry); 18855 } 18856 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom(byte[] data) 18857 throws com.google.protobuf.InvalidProtocolBufferException { 18858 return PARSER.parseFrom(data); 18859 } 18860 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18861 byte[] data, 18862 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18863 throws com.google.protobuf.InvalidProtocolBufferException { 18864 return PARSER.parseFrom(data, extensionRegistry); 18865 } 18866 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom(java.io.InputStream input) 18867 throws java.io.IOException { 18868 return PARSER.parseFrom(input); 18869 } 18870 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18871 java.io.InputStream input, 18872 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18873 throws java.io.IOException { 18874 return PARSER.parseFrom(input, extensionRegistry); 18875 } 18876 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseDelimitedFrom(java.io.InputStream input) 18877 throws java.io.IOException { 18878 return PARSER.parseDelimitedFrom(input); 18879 } 18880 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseDelimitedFrom( 18881 java.io.InputStream input, 18882 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18883 throws java.io.IOException { 18884 return PARSER.parseDelimitedFrom(input, extensionRegistry); 18885 } 18886 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18887 com.google.protobuf.CodedInputStream input) 18888 throws java.io.IOException { 18889 return PARSER.parseFrom(input); 18890 } 18891 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parseFrom( 18892 com.google.protobuf.CodedInputStream input, 18893 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 18894 throws java.io.IOException { 18895 return PARSER.parseFrom(input, extensionRegistry); 18896 } 18897 18898 public static Builder newBuilder() { return Builder.create(); } 18899 public Builder newBuilderForType() { return newBuilder(); } 18900 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry prototype) { 18901 return newBuilder().mergeFrom(prototype); 18902 } 18903 public Builder toBuilder() { return newBuilder(this); } 18904 18905 @java.lang.Override 18906 protected Builder newBuilderForType( 18907 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18908 Builder builder = new Builder(parent); 18909 return builder; 18910 } 18911 /** 18912 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection.Entry} 18913 */ 18914 public static final class Builder extends 18915 com.google.protobuf.GeneratedMessage.Builder<Builder> 18916 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.EntryOrBuilder { 18917 public static final com.google.protobuf.Descriptors.Descriptor 18918 getDescriptor() { 18919 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 18920 } 18921 18922 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 18923 internalGetFieldAccessorTable() { 18924 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable 18925 .ensureFieldAccessorsInitialized( 18926 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.Builder.class); 18927 } 18928 18929 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.newBuilder() 18930 private Builder() { 18931 maybeForceBuilderInitialization(); 18932 } 18933 18934 private Builder( 18935 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 18936 super(parent); 18937 maybeForceBuilderInitialization(); 18938 } 18939 private void maybeForceBuilderInitialization() { 18940 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 18941 } 18942 } 18943 private static Builder create() { 18944 return new Builder(); 18945 } 18946 18947 public Builder clear() { 18948 super.clear(); 18949 id_ = 0; 18950 bitField0_ = (bitField0_ & ~0x00000001); 18951 str_ = ""; 18952 bitField0_ = (bitField0_ & ~0x00000002); 18953 return this; 18954 } 18955 18956 public Builder clone() { 18957 return create().mergeFrom(buildPartial()); 18958 } 18959 18960 public com.google.protobuf.Descriptors.Descriptor 18961 getDescriptorForType() { 18962 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 18963 } 18964 18965 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry getDefaultInstanceForType() { 18966 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.getDefaultInstance(); 18967 } 18968 18969 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry build() { 18970 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry result = buildPartial(); 18971 if (!result.isInitialized()) { 18972 throw newUninitializedMessageException(result); 18973 } 18974 return result; 18975 } 18976 18977 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry buildPartial() { 18978 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry(this); 18979 int from_bitField0_ = bitField0_; 18980 int to_bitField0_ = 0; 18981 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 18982 to_bitField0_ |= 0x00000001; 18983 } 18984 result.id_ = id_; 18985 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 18986 to_bitField0_ |= 0x00000002; 18987 } 18988 result.str_ = str_; 18989 result.bitField0_ = to_bitField0_; 18990 onBuilt(); 18991 return result; 18992 } 18993 18994 public Builder mergeFrom(com.google.protobuf.Message other) { 18995 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry) { 18996 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry)other); 18997 } else { 18998 super.mergeFrom(other); 18999 return this; 19000 } 19001 } 19002 19003 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry other) { 19004 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry.getDefaultInstance()) return this; 19005 if (other.hasId()) { 19006 setId(other.getId()); 19007 } 19008 if (other.hasStr()) { 19009 bitField0_ |= 0x00000002; 19010 str_ = other.str_; 19011 onChanged(); 19012 } 19013 this.mergeUnknownFields(other.getUnknownFields()); 19014 return this; 19015 } 19016 19017 public final boolean isInitialized() { 19018 return true; 19019 } 19020 19021 public Builder mergeFrom( 19022 com.google.protobuf.CodedInputStream input, 19023 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19024 throws java.io.IOException { 19025 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry parsedMessage = null; 19026 try { 19027 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 19028 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 19029 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Entry) e.getUnfinishedMessage(); 19030 throw e; 19031 } finally { 19032 if (parsedMessage != null) { 19033 mergeFrom(parsedMessage); 19034 } 19035 } 19036 return this; 19037 } 19038 private int bitField0_; 19039 19040 // optional uint32 id = 1; 19041 private int id_ ; 19042 /** 19043 * <code>optional uint32 id = 1;</code> 19044 */ 19045 public boolean hasId() { 19046 return ((bitField0_ & 0x00000001) == 0x00000001); 19047 } 19048 /** 19049 * <code>optional uint32 id = 1;</code> 19050 */ 19051 public int getId() { 19052 return id_; 19053 } 19054 /** 19055 * <code>optional uint32 id = 1;</code> 19056 */ 19057 public Builder setId(int value) { 19058 bitField0_ |= 0x00000001; 19059 id_ = value; 19060 onChanged(); 19061 return this; 19062 } 19063 /** 19064 * <code>optional uint32 id = 1;</code> 19065 */ 19066 public Builder clearId() { 19067 bitField0_ = (bitField0_ & ~0x00000001); 19068 id_ = 0; 19069 onChanged(); 19070 return this; 19071 } 19072 19073 // optional string str = 2; 19074 private java.lang.Object str_ = ""; 19075 /** 19076 * <code>optional string str = 2;</code> 19077 */ 19078 public boolean hasStr() { 19079 return ((bitField0_ & 0x00000002) == 0x00000002); 19080 } 19081 /** 19082 * <code>optional string str = 2;</code> 19083 */ 19084 public java.lang.String getStr() { 19085 java.lang.Object ref = str_; 19086 if (!(ref instanceof java.lang.String)) { 19087 java.lang.String s = ((com.google.protobuf.ByteString) ref) 19088 .toStringUtf8(); 19089 str_ = s; 19090 return s; 19091 } else { 19092 return (java.lang.String) ref; 19093 } 19094 } 19095 /** 19096 * <code>optional string str = 2;</code> 19097 */ 19098 public com.google.protobuf.ByteString 19099 getStrBytes() { 19100 java.lang.Object ref = str_; 19101 if (ref instanceof String) { 19102 com.google.protobuf.ByteString b = 19103 com.google.protobuf.ByteString.copyFromUtf8( 19104 (java.lang.String) ref); 19105 str_ = b; 19106 return b; 19107 } else { 19108 return (com.google.protobuf.ByteString) ref; 19109 } 19110 } 19111 /** 19112 * <code>optional string str = 2;</code> 19113 */ 19114 public Builder setStr( 19115 java.lang.String value) { 19116 if (value == null) { 19117 throw new NullPointerException(); 19118 } 19119 bitField0_ |= 0x00000002; 19120 str_ = value; 19121 onChanged(); 19122 return this; 19123 } 19124 /** 19125 * <code>optional string str = 2;</code> 19126 */ 19127 public Builder clearStr() { 19128 bitField0_ = (bitField0_ & ~0x00000002); 19129 str_ = getDefaultInstance().getStr(); 19130 onChanged(); 19131 return this; 19132 } 19133 /** 19134 * <code>optional string str = 2;</code> 19135 */ 19136 public Builder setStrBytes( 19137 com.google.protobuf.ByteString value) { 19138 if (value == null) { 19139 throw new NullPointerException(); 19140 } 19141 bitField0_ |= 0x00000002; 19142 str_ = value; 19143 onChanged(); 19144 return this; 19145 } 19146 19147 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.StringTableSection.Entry) 19148 } 19149 19150 static { 19151 defaultInstance = new Entry(true); 19152 defaultInstance.initFields(); 19153 } 19154 19155 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.StringTableSection.Entry) 19156 } 19157 19158 private int bitField0_; 19159 // optional uint32 numEntry = 1; 19160 public static final int NUMENTRY_FIELD_NUMBER = 1; 19161 private int numEntry_; 19162 /** 19163 * <code>optional uint32 numEntry = 1;</code> 19164 * 19165 * <pre> 19166 * repeated Entry 19167 * </pre> 19168 */ 19169 public boolean hasNumEntry() { 19170 return ((bitField0_ & 0x00000001) == 0x00000001); 19171 } 19172 /** 19173 * <code>optional uint32 numEntry = 1;</code> 19174 * 19175 * <pre> 19176 * repeated Entry 19177 * </pre> 19178 */ 19179 public int getNumEntry() { 19180 return numEntry_; 19181 } 19182 19183 private void initFields() { 19184 numEntry_ = 0; 19185 } 19186 private byte memoizedIsInitialized = -1; 19187 public final boolean isInitialized() { 19188 byte isInitialized = memoizedIsInitialized; 19189 if (isInitialized != -1) return isInitialized == 1; 19190 19191 memoizedIsInitialized = 1; 19192 return true; 19193 } 19194 19195 public void writeTo(com.google.protobuf.CodedOutputStream output) 19196 throws java.io.IOException { 19197 getSerializedSize(); 19198 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19199 output.writeUInt32(1, numEntry_); 19200 } 19201 getUnknownFields().writeTo(output); 19202 } 19203 19204 private int memoizedSerializedSize = -1; 19205 public int getSerializedSize() { 19206 int size = memoizedSerializedSize; 19207 if (size != -1) return size; 19208 19209 size = 0; 19210 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19211 size += com.google.protobuf.CodedOutputStream 19212 .computeUInt32Size(1, numEntry_); 19213 } 19214 size += getUnknownFields().getSerializedSize(); 19215 memoizedSerializedSize = size; 19216 return size; 19217 } 19218 19219 private static final long serialVersionUID = 0L; 19220 @java.lang.Override 19221 protected java.lang.Object writeReplace() 19222 throws java.io.ObjectStreamException { 19223 return super.writeReplace(); 19224 } 19225 19226 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19227 com.google.protobuf.ByteString data) 19228 throws com.google.protobuf.InvalidProtocolBufferException { 19229 return PARSER.parseFrom(data); 19230 } 19231 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19232 com.google.protobuf.ByteString data, 19233 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19234 throws com.google.protobuf.InvalidProtocolBufferException { 19235 return PARSER.parseFrom(data, extensionRegistry); 19236 } 19237 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom(byte[] data) 19238 throws com.google.protobuf.InvalidProtocolBufferException { 19239 return PARSER.parseFrom(data); 19240 } 19241 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19242 byte[] data, 19243 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19244 throws com.google.protobuf.InvalidProtocolBufferException { 19245 return PARSER.parseFrom(data, extensionRegistry); 19246 } 19247 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom(java.io.InputStream input) 19248 throws java.io.IOException { 19249 return PARSER.parseFrom(input); 19250 } 19251 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19252 java.io.InputStream input, 19253 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19254 throws java.io.IOException { 19255 return PARSER.parseFrom(input, extensionRegistry); 19256 } 19257 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseDelimitedFrom(java.io.InputStream input) 19258 throws java.io.IOException { 19259 return PARSER.parseDelimitedFrom(input); 19260 } 19261 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseDelimitedFrom( 19262 java.io.InputStream input, 19263 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19264 throws java.io.IOException { 19265 return PARSER.parseDelimitedFrom(input, extensionRegistry); 19266 } 19267 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19268 com.google.protobuf.CodedInputStream input) 19269 throws java.io.IOException { 19270 return PARSER.parseFrom(input); 19271 } 19272 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parseFrom( 19273 com.google.protobuf.CodedInputStream input, 19274 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19275 throws java.io.IOException { 19276 return PARSER.parseFrom(input, extensionRegistry); 19277 } 19278 19279 public static Builder newBuilder() { return Builder.create(); } 19280 public Builder newBuilderForType() { return newBuilder(); } 19281 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection prototype) { 19282 return newBuilder().mergeFrom(prototype); 19283 } 19284 public Builder toBuilder() { return newBuilder(this); } 19285 19286 @java.lang.Override 19287 protected Builder newBuilderForType( 19288 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19289 Builder builder = new Builder(parent); 19290 return builder; 19291 } 19292 /** 19293 * Protobuf type {@code hadoop.hdfs.fsimage.StringTableSection} 19294 * 19295 * <pre> 19296 ** 19297 * This section maps string to id 19298 * NAME: STRING_TABLE 19299 * </pre> 19300 */ 19301 public static final class Builder extends 19302 com.google.protobuf.GeneratedMessage.Builder<Builder> 19303 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSectionOrBuilder { 19304 public static final com.google.protobuf.Descriptors.Descriptor 19305 getDescriptor() { 19306 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 19307 } 19308 19309 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19310 internalGetFieldAccessorTable() { 19311 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable 19312 .ensureFieldAccessorsInitialized( 19313 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.Builder.class); 19314 } 19315 19316 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.newBuilder() 19317 private Builder() { 19318 maybeForceBuilderInitialization(); 19319 } 19320 19321 private Builder( 19322 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19323 super(parent); 19324 maybeForceBuilderInitialization(); 19325 } 19326 private void maybeForceBuilderInitialization() { 19327 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 19328 } 19329 } 19330 private static Builder create() { 19331 return new Builder(); 19332 } 19333 19334 public Builder clear() { 19335 super.clear(); 19336 numEntry_ = 0; 19337 bitField0_ = (bitField0_ & ~0x00000001); 19338 return this; 19339 } 19340 19341 public Builder clone() { 19342 return create().mergeFrom(buildPartial()); 19343 } 19344 19345 public com.google.protobuf.Descriptors.Descriptor 19346 getDescriptorForType() { 19347 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 19348 } 19349 19350 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection getDefaultInstanceForType() { 19351 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.getDefaultInstance(); 19352 } 19353 19354 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection build() { 19355 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection result = buildPartial(); 19356 if (!result.isInitialized()) { 19357 throw newUninitializedMessageException(result); 19358 } 19359 return result; 19360 } 19361 19362 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection buildPartial() { 19363 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection(this); 19364 int from_bitField0_ = bitField0_; 19365 int to_bitField0_ = 0; 19366 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 19367 to_bitField0_ |= 0x00000001; 19368 } 19369 result.numEntry_ = numEntry_; 19370 result.bitField0_ = to_bitField0_; 19371 onBuilt(); 19372 return result; 19373 } 19374 19375 public Builder mergeFrom(com.google.protobuf.Message other) { 19376 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection) { 19377 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection)other); 19378 } else { 19379 super.mergeFrom(other); 19380 return this; 19381 } 19382 } 19383 19384 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection other) { 19385 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection.getDefaultInstance()) return this; 19386 if (other.hasNumEntry()) { 19387 setNumEntry(other.getNumEntry()); 19388 } 19389 this.mergeUnknownFields(other.getUnknownFields()); 19390 return this; 19391 } 19392 19393 public final boolean isInitialized() { 19394 return true; 19395 } 19396 19397 public Builder mergeFrom( 19398 com.google.protobuf.CodedInputStream input, 19399 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19400 throws java.io.IOException { 19401 org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection parsedMessage = null; 19402 try { 19403 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 19404 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 19405 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.StringTableSection) e.getUnfinishedMessage(); 19406 throw e; 19407 } finally { 19408 if (parsedMessage != null) { 19409 mergeFrom(parsedMessage); 19410 } 19411 } 19412 return this; 19413 } 19414 private int bitField0_; 19415 19416 // optional uint32 numEntry = 1; 19417 private int numEntry_ ; 19418 /** 19419 * <code>optional uint32 numEntry = 1;</code> 19420 * 19421 * <pre> 19422 * repeated Entry 19423 * </pre> 19424 */ 19425 public boolean hasNumEntry() { 19426 return ((bitField0_ & 0x00000001) == 0x00000001); 19427 } 19428 /** 19429 * <code>optional uint32 numEntry = 1;</code> 19430 * 19431 * <pre> 19432 * repeated Entry 19433 * </pre> 19434 */ 19435 public int getNumEntry() { 19436 return numEntry_; 19437 } 19438 /** 19439 * <code>optional uint32 numEntry = 1;</code> 19440 * 19441 * <pre> 19442 * repeated Entry 19443 * </pre> 19444 */ 19445 public Builder setNumEntry(int value) { 19446 bitField0_ |= 0x00000001; 19447 numEntry_ = value; 19448 onChanged(); 19449 return this; 19450 } 19451 /** 19452 * <code>optional uint32 numEntry = 1;</code> 19453 * 19454 * <pre> 19455 * repeated Entry 19456 * </pre> 19457 */ 19458 public Builder clearNumEntry() { 19459 bitField0_ = (bitField0_ & ~0x00000001); 19460 numEntry_ = 0; 19461 onChanged(); 19462 return this; 19463 } 19464 19465 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.StringTableSection) 19466 } 19467 19468 static { 19469 defaultInstance = new StringTableSection(true); 19470 defaultInstance.initFields(); 19471 } 19472 19473 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.StringTableSection) 19474 } 19475 19476 public interface SecretManagerSectionOrBuilder 19477 extends com.google.protobuf.MessageOrBuilder { 19478 19479 // optional uint32 currentId = 1; 19480 /** 19481 * <code>optional uint32 currentId = 1;</code> 19482 */ 19483 boolean hasCurrentId(); 19484 /** 19485 * <code>optional uint32 currentId = 1;</code> 19486 */ 19487 int getCurrentId(); 19488 19489 // optional uint32 tokenSequenceNumber = 2; 19490 /** 19491 * <code>optional uint32 tokenSequenceNumber = 2;</code> 19492 */ 19493 boolean hasTokenSequenceNumber(); 19494 /** 19495 * <code>optional uint32 tokenSequenceNumber = 2;</code> 19496 */ 19497 int getTokenSequenceNumber(); 19498 19499 // optional uint32 numKeys = 3; 19500 /** 19501 * <code>optional uint32 numKeys = 3;</code> 19502 */ 19503 boolean hasNumKeys(); 19504 /** 19505 * <code>optional uint32 numKeys = 3;</code> 19506 */ 19507 int getNumKeys(); 19508 19509 // optional uint32 numTokens = 4; 19510 /** 19511 * <code>optional uint32 numTokens = 4;</code> 19512 * 19513 * <pre> 19514 * repeated DelegationKey keys 19515 * repeated PersistToken tokens 19516 * </pre> 19517 */ 19518 boolean hasNumTokens(); 19519 /** 19520 * <code>optional uint32 numTokens = 4;</code> 19521 * 19522 * <pre> 19523 * repeated DelegationKey keys 19524 * repeated PersistToken tokens 19525 * </pre> 19526 */ 19527 int getNumTokens(); 19528 } 19529 /** 19530 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection} 19531 */ 19532 public static final class SecretManagerSection extends 19533 com.google.protobuf.GeneratedMessage 19534 implements SecretManagerSectionOrBuilder { 19535 // Use SecretManagerSection.newBuilder() to construct. 19536 private SecretManagerSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 19537 super(builder); 19538 this.unknownFields = builder.getUnknownFields(); 19539 } 19540 private SecretManagerSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 19541 19542 private static final SecretManagerSection defaultInstance; 19543 public static SecretManagerSection getDefaultInstance() { 19544 return defaultInstance; 19545 } 19546 19547 public SecretManagerSection getDefaultInstanceForType() { 19548 return defaultInstance; 19549 } 19550 19551 private final com.google.protobuf.UnknownFieldSet unknownFields; 19552 @java.lang.Override 19553 public final com.google.protobuf.UnknownFieldSet 19554 getUnknownFields() { 19555 return this.unknownFields; 19556 } 19557 private SecretManagerSection( 19558 com.google.protobuf.CodedInputStream input, 19559 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19560 throws com.google.protobuf.InvalidProtocolBufferException { 19561 initFields(); 19562 int mutable_bitField0_ = 0; 19563 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 19564 com.google.protobuf.UnknownFieldSet.newBuilder(); 19565 try { 19566 boolean done = false; 19567 while (!done) { 19568 int tag = input.readTag(); 19569 switch (tag) { 19570 case 0: 19571 done = true; 19572 break; 19573 default: { 19574 if (!parseUnknownField(input, unknownFields, 19575 extensionRegistry, tag)) { 19576 done = true; 19577 } 19578 break; 19579 } 19580 case 8: { 19581 bitField0_ |= 0x00000001; 19582 currentId_ = input.readUInt32(); 19583 break; 19584 } 19585 case 16: { 19586 bitField0_ |= 0x00000002; 19587 tokenSequenceNumber_ = input.readUInt32(); 19588 break; 19589 } 19590 case 24: { 19591 bitField0_ |= 0x00000004; 19592 numKeys_ = input.readUInt32(); 19593 break; 19594 } 19595 case 32: { 19596 bitField0_ |= 0x00000008; 19597 numTokens_ = input.readUInt32(); 19598 break; 19599 } 19600 } 19601 } 19602 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 19603 throw e.setUnfinishedMessage(this); 19604 } catch (java.io.IOException e) { 19605 throw new com.google.protobuf.InvalidProtocolBufferException( 19606 e.getMessage()).setUnfinishedMessage(this); 19607 } finally { 19608 this.unknownFields = unknownFields.build(); 19609 makeExtensionsImmutable(); 19610 } 19611 } 19612 public static final com.google.protobuf.Descriptors.Descriptor 19613 getDescriptor() { 19614 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 19615 } 19616 19617 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19618 internalGetFieldAccessorTable() { 19619 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable 19620 .ensureFieldAccessorsInitialized( 19621 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.Builder.class); 19622 } 19623 19624 public static com.google.protobuf.Parser<SecretManagerSection> PARSER = 19625 new com.google.protobuf.AbstractParser<SecretManagerSection>() { 19626 public SecretManagerSection parsePartialFrom( 19627 com.google.protobuf.CodedInputStream input, 19628 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19629 throws com.google.protobuf.InvalidProtocolBufferException { 19630 return new SecretManagerSection(input, extensionRegistry); 19631 } 19632 }; 19633 19634 @java.lang.Override 19635 public com.google.protobuf.Parser<SecretManagerSection> getParserForType() { 19636 return PARSER; 19637 } 19638 19639 public interface DelegationKeyOrBuilder 19640 extends com.google.protobuf.MessageOrBuilder { 19641 19642 // optional uint32 id = 1; 19643 /** 19644 * <code>optional uint32 id = 1;</code> 19645 */ 19646 boolean hasId(); 19647 /** 19648 * <code>optional uint32 id = 1;</code> 19649 */ 19650 int getId(); 19651 19652 // optional uint64 expiryDate = 2; 19653 /** 19654 * <code>optional uint64 expiryDate = 2;</code> 19655 */ 19656 boolean hasExpiryDate(); 19657 /** 19658 * <code>optional uint64 expiryDate = 2;</code> 19659 */ 19660 long getExpiryDate(); 19661 19662 // optional bytes key = 3; 19663 /** 19664 * <code>optional bytes key = 3;</code> 19665 */ 19666 boolean hasKey(); 19667 /** 19668 * <code>optional bytes key = 3;</code> 19669 */ 19670 com.google.protobuf.ByteString getKey(); 19671 } 19672 /** 19673 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey} 19674 */ 19675 public static final class DelegationKey extends 19676 com.google.protobuf.GeneratedMessage 19677 implements DelegationKeyOrBuilder { 19678 // Use DelegationKey.newBuilder() to construct. 19679 private DelegationKey(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 19680 super(builder); 19681 this.unknownFields = builder.getUnknownFields(); 19682 } 19683 private DelegationKey(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 19684 19685 private static final DelegationKey defaultInstance; 19686 public static DelegationKey getDefaultInstance() { 19687 return defaultInstance; 19688 } 19689 19690 public DelegationKey getDefaultInstanceForType() { 19691 return defaultInstance; 19692 } 19693 19694 private final com.google.protobuf.UnknownFieldSet unknownFields; 19695 @java.lang.Override 19696 public final com.google.protobuf.UnknownFieldSet 19697 getUnknownFields() { 19698 return this.unknownFields; 19699 } 19700 private DelegationKey( 19701 com.google.protobuf.CodedInputStream input, 19702 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19703 throws com.google.protobuf.InvalidProtocolBufferException { 19704 initFields(); 19705 int mutable_bitField0_ = 0; 19706 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 19707 com.google.protobuf.UnknownFieldSet.newBuilder(); 19708 try { 19709 boolean done = false; 19710 while (!done) { 19711 int tag = input.readTag(); 19712 switch (tag) { 19713 case 0: 19714 done = true; 19715 break; 19716 default: { 19717 if (!parseUnknownField(input, unknownFields, 19718 extensionRegistry, tag)) { 19719 done = true; 19720 } 19721 break; 19722 } 19723 case 8: { 19724 bitField0_ |= 0x00000001; 19725 id_ = input.readUInt32(); 19726 break; 19727 } 19728 case 16: { 19729 bitField0_ |= 0x00000002; 19730 expiryDate_ = input.readUInt64(); 19731 break; 19732 } 19733 case 26: { 19734 bitField0_ |= 0x00000004; 19735 key_ = input.readBytes(); 19736 break; 19737 } 19738 } 19739 } 19740 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 19741 throw e.setUnfinishedMessage(this); 19742 } catch (java.io.IOException e) { 19743 throw new com.google.protobuf.InvalidProtocolBufferException( 19744 e.getMessage()).setUnfinishedMessage(this); 19745 } finally { 19746 this.unknownFields = unknownFields.build(); 19747 makeExtensionsImmutable(); 19748 } 19749 } 19750 public static final com.google.protobuf.Descriptors.Descriptor 19751 getDescriptor() { 19752 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 19753 } 19754 19755 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19756 internalGetFieldAccessorTable() { 19757 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable 19758 .ensureFieldAccessorsInitialized( 19759 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.Builder.class); 19760 } 19761 19762 public static com.google.protobuf.Parser<DelegationKey> PARSER = 19763 new com.google.protobuf.AbstractParser<DelegationKey>() { 19764 public DelegationKey parsePartialFrom( 19765 com.google.protobuf.CodedInputStream input, 19766 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19767 throws com.google.protobuf.InvalidProtocolBufferException { 19768 return new DelegationKey(input, extensionRegistry); 19769 } 19770 }; 19771 19772 @java.lang.Override 19773 public com.google.protobuf.Parser<DelegationKey> getParserForType() { 19774 return PARSER; 19775 } 19776 19777 private int bitField0_; 19778 // optional uint32 id = 1; 19779 public static final int ID_FIELD_NUMBER = 1; 19780 private int id_; 19781 /** 19782 * <code>optional uint32 id = 1;</code> 19783 */ 19784 public boolean hasId() { 19785 return ((bitField0_ & 0x00000001) == 0x00000001); 19786 } 19787 /** 19788 * <code>optional uint32 id = 1;</code> 19789 */ 19790 public int getId() { 19791 return id_; 19792 } 19793 19794 // optional uint64 expiryDate = 2; 19795 public static final int EXPIRYDATE_FIELD_NUMBER = 2; 19796 private long expiryDate_; 19797 /** 19798 * <code>optional uint64 expiryDate = 2;</code> 19799 */ 19800 public boolean hasExpiryDate() { 19801 return ((bitField0_ & 0x00000002) == 0x00000002); 19802 } 19803 /** 19804 * <code>optional uint64 expiryDate = 2;</code> 19805 */ 19806 public long getExpiryDate() { 19807 return expiryDate_; 19808 } 19809 19810 // optional bytes key = 3; 19811 public static final int KEY_FIELD_NUMBER = 3; 19812 private com.google.protobuf.ByteString key_; 19813 /** 19814 * <code>optional bytes key = 3;</code> 19815 */ 19816 public boolean hasKey() { 19817 return ((bitField0_ & 0x00000004) == 0x00000004); 19818 } 19819 /** 19820 * <code>optional bytes key = 3;</code> 19821 */ 19822 public com.google.protobuf.ByteString getKey() { 19823 return key_; 19824 } 19825 19826 private void initFields() { 19827 id_ = 0; 19828 expiryDate_ = 0L; 19829 key_ = com.google.protobuf.ByteString.EMPTY; 19830 } 19831 private byte memoizedIsInitialized = -1; 19832 public final boolean isInitialized() { 19833 byte isInitialized = memoizedIsInitialized; 19834 if (isInitialized != -1) return isInitialized == 1; 19835 19836 memoizedIsInitialized = 1; 19837 return true; 19838 } 19839 19840 public void writeTo(com.google.protobuf.CodedOutputStream output) 19841 throws java.io.IOException { 19842 getSerializedSize(); 19843 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19844 output.writeUInt32(1, id_); 19845 } 19846 if (((bitField0_ & 0x00000002) == 0x00000002)) { 19847 output.writeUInt64(2, expiryDate_); 19848 } 19849 if (((bitField0_ & 0x00000004) == 0x00000004)) { 19850 output.writeBytes(3, key_); 19851 } 19852 getUnknownFields().writeTo(output); 19853 } 19854 19855 private int memoizedSerializedSize = -1; 19856 public int getSerializedSize() { 19857 int size = memoizedSerializedSize; 19858 if (size != -1) return size; 19859 19860 size = 0; 19861 if (((bitField0_ & 0x00000001) == 0x00000001)) { 19862 size += com.google.protobuf.CodedOutputStream 19863 .computeUInt32Size(1, id_); 19864 } 19865 if (((bitField0_ & 0x00000002) == 0x00000002)) { 19866 size += com.google.protobuf.CodedOutputStream 19867 .computeUInt64Size(2, expiryDate_); 19868 } 19869 if (((bitField0_ & 0x00000004) == 0x00000004)) { 19870 size += com.google.protobuf.CodedOutputStream 19871 .computeBytesSize(3, key_); 19872 } 19873 size += getUnknownFields().getSerializedSize(); 19874 memoizedSerializedSize = size; 19875 return size; 19876 } 19877 19878 private static final long serialVersionUID = 0L; 19879 @java.lang.Override 19880 protected java.lang.Object writeReplace() 19881 throws java.io.ObjectStreamException { 19882 return super.writeReplace(); 19883 } 19884 19885 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19886 com.google.protobuf.ByteString data) 19887 throws com.google.protobuf.InvalidProtocolBufferException { 19888 return PARSER.parseFrom(data); 19889 } 19890 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19891 com.google.protobuf.ByteString data, 19892 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19893 throws com.google.protobuf.InvalidProtocolBufferException { 19894 return PARSER.parseFrom(data, extensionRegistry); 19895 } 19896 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom(byte[] data) 19897 throws com.google.protobuf.InvalidProtocolBufferException { 19898 return PARSER.parseFrom(data); 19899 } 19900 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19901 byte[] data, 19902 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19903 throws com.google.protobuf.InvalidProtocolBufferException { 19904 return PARSER.parseFrom(data, extensionRegistry); 19905 } 19906 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom(java.io.InputStream input) 19907 throws java.io.IOException { 19908 return PARSER.parseFrom(input); 19909 } 19910 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19911 java.io.InputStream input, 19912 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19913 throws java.io.IOException { 19914 return PARSER.parseFrom(input, extensionRegistry); 19915 } 19916 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseDelimitedFrom(java.io.InputStream input) 19917 throws java.io.IOException { 19918 return PARSER.parseDelimitedFrom(input); 19919 } 19920 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseDelimitedFrom( 19921 java.io.InputStream input, 19922 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19923 throws java.io.IOException { 19924 return PARSER.parseDelimitedFrom(input, extensionRegistry); 19925 } 19926 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19927 com.google.protobuf.CodedInputStream input) 19928 throws java.io.IOException { 19929 return PARSER.parseFrom(input); 19930 } 19931 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parseFrom( 19932 com.google.protobuf.CodedInputStream input, 19933 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 19934 throws java.io.IOException { 19935 return PARSER.parseFrom(input, extensionRegistry); 19936 } 19937 19938 public static Builder newBuilder() { return Builder.create(); } 19939 public Builder newBuilderForType() { return newBuilder(); } 19940 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey prototype) { 19941 return newBuilder().mergeFrom(prototype); 19942 } 19943 public Builder toBuilder() { return newBuilder(this); } 19944 19945 @java.lang.Override 19946 protected Builder newBuilderForType( 19947 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19948 Builder builder = new Builder(parent); 19949 return builder; 19950 } 19951 /** 19952 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey} 19953 */ 19954 public static final class Builder extends 19955 com.google.protobuf.GeneratedMessage.Builder<Builder> 19956 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKeyOrBuilder { 19957 public static final com.google.protobuf.Descriptors.Descriptor 19958 getDescriptor() { 19959 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 19960 } 19961 19962 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 19963 internalGetFieldAccessorTable() { 19964 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable 19965 .ensureFieldAccessorsInitialized( 19966 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.Builder.class); 19967 } 19968 19969 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.newBuilder() 19970 private Builder() { 19971 maybeForceBuilderInitialization(); 19972 } 19973 19974 private Builder( 19975 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 19976 super(parent); 19977 maybeForceBuilderInitialization(); 19978 } 19979 private void maybeForceBuilderInitialization() { 19980 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 19981 } 19982 } 19983 private static Builder create() { 19984 return new Builder(); 19985 } 19986 19987 public Builder clear() { 19988 super.clear(); 19989 id_ = 0; 19990 bitField0_ = (bitField0_ & ~0x00000001); 19991 expiryDate_ = 0L; 19992 bitField0_ = (bitField0_ & ~0x00000002); 19993 key_ = com.google.protobuf.ByteString.EMPTY; 19994 bitField0_ = (bitField0_ & ~0x00000004); 19995 return this; 19996 } 19997 19998 public Builder clone() { 19999 return create().mergeFrom(buildPartial()); 20000 } 20001 20002 public com.google.protobuf.Descriptors.Descriptor 20003 getDescriptorForType() { 20004 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 20005 } 20006 20007 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey getDefaultInstanceForType() { 20008 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.getDefaultInstance(); 20009 } 20010 20011 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey build() { 20012 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey result = buildPartial(); 20013 if (!result.isInitialized()) { 20014 throw newUninitializedMessageException(result); 20015 } 20016 return result; 20017 } 20018 20019 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey buildPartial() { 20020 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey(this); 20021 int from_bitField0_ = bitField0_; 20022 int to_bitField0_ = 0; 20023 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 20024 to_bitField0_ |= 0x00000001; 20025 } 20026 result.id_ = id_; 20027 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 20028 to_bitField0_ |= 0x00000002; 20029 } 20030 result.expiryDate_ = expiryDate_; 20031 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 20032 to_bitField0_ |= 0x00000004; 20033 } 20034 result.key_ = key_; 20035 result.bitField0_ = to_bitField0_; 20036 onBuilt(); 20037 return result; 20038 } 20039 20040 public Builder mergeFrom(com.google.protobuf.Message other) { 20041 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey) { 20042 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey)other); 20043 } else { 20044 super.mergeFrom(other); 20045 return this; 20046 } 20047 } 20048 20049 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey other) { 20050 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey.getDefaultInstance()) return this; 20051 if (other.hasId()) { 20052 setId(other.getId()); 20053 } 20054 if (other.hasExpiryDate()) { 20055 setExpiryDate(other.getExpiryDate()); 20056 } 20057 if (other.hasKey()) { 20058 setKey(other.getKey()); 20059 } 20060 this.mergeUnknownFields(other.getUnknownFields()); 20061 return this; 20062 } 20063 20064 public final boolean isInitialized() { 20065 return true; 20066 } 20067 20068 public Builder mergeFrom( 20069 com.google.protobuf.CodedInputStream input, 20070 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20071 throws java.io.IOException { 20072 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey parsedMessage = null; 20073 try { 20074 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 20075 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 20076 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.DelegationKey) e.getUnfinishedMessage(); 20077 throw e; 20078 } finally { 20079 if (parsedMessage != null) { 20080 mergeFrom(parsedMessage); 20081 } 20082 } 20083 return this; 20084 } 20085 private int bitField0_; 20086 20087 // optional uint32 id = 1; 20088 private int id_ ; 20089 /** 20090 * <code>optional uint32 id = 1;</code> 20091 */ 20092 public boolean hasId() { 20093 return ((bitField0_ & 0x00000001) == 0x00000001); 20094 } 20095 /** 20096 * <code>optional uint32 id = 1;</code> 20097 */ 20098 public int getId() { 20099 return id_; 20100 } 20101 /** 20102 * <code>optional uint32 id = 1;</code> 20103 */ 20104 public Builder setId(int value) { 20105 bitField0_ |= 0x00000001; 20106 id_ = value; 20107 onChanged(); 20108 return this; 20109 } 20110 /** 20111 * <code>optional uint32 id = 1;</code> 20112 */ 20113 public Builder clearId() { 20114 bitField0_ = (bitField0_ & ~0x00000001); 20115 id_ = 0; 20116 onChanged(); 20117 return this; 20118 } 20119 20120 // optional uint64 expiryDate = 2; 20121 private long expiryDate_ ; 20122 /** 20123 * <code>optional uint64 expiryDate = 2;</code> 20124 */ 20125 public boolean hasExpiryDate() { 20126 return ((bitField0_ & 0x00000002) == 0x00000002); 20127 } 20128 /** 20129 * <code>optional uint64 expiryDate = 2;</code> 20130 */ 20131 public long getExpiryDate() { 20132 return expiryDate_; 20133 } 20134 /** 20135 * <code>optional uint64 expiryDate = 2;</code> 20136 */ 20137 public Builder setExpiryDate(long value) { 20138 bitField0_ |= 0x00000002; 20139 expiryDate_ = value; 20140 onChanged(); 20141 return this; 20142 } 20143 /** 20144 * <code>optional uint64 expiryDate = 2;</code> 20145 */ 20146 public Builder clearExpiryDate() { 20147 bitField0_ = (bitField0_ & ~0x00000002); 20148 expiryDate_ = 0L; 20149 onChanged(); 20150 return this; 20151 } 20152 20153 // optional bytes key = 3; 20154 private com.google.protobuf.ByteString key_ = com.google.protobuf.ByteString.EMPTY; 20155 /** 20156 * <code>optional bytes key = 3;</code> 20157 */ 20158 public boolean hasKey() { 20159 return ((bitField0_ & 0x00000004) == 0x00000004); 20160 } 20161 /** 20162 * <code>optional bytes key = 3;</code> 20163 */ 20164 public com.google.protobuf.ByteString getKey() { 20165 return key_; 20166 } 20167 /** 20168 * <code>optional bytes key = 3;</code> 20169 */ 20170 public Builder setKey(com.google.protobuf.ByteString value) { 20171 if (value == null) { 20172 throw new NullPointerException(); 20173 } 20174 bitField0_ |= 0x00000004; 20175 key_ = value; 20176 onChanged(); 20177 return this; 20178 } 20179 /** 20180 * <code>optional bytes key = 3;</code> 20181 */ 20182 public Builder clearKey() { 20183 bitField0_ = (bitField0_ & ~0x00000004); 20184 key_ = getDefaultInstance().getKey(); 20185 onChanged(); 20186 return this; 20187 } 20188 20189 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey) 20190 } 20191 20192 static { 20193 defaultInstance = new DelegationKey(true); 20194 defaultInstance.initFields(); 20195 } 20196 20197 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection.DelegationKey) 20198 } 20199 20200 public interface PersistTokenOrBuilder 20201 extends com.google.protobuf.MessageOrBuilder { 20202 20203 // optional uint32 version = 1; 20204 /** 20205 * <code>optional uint32 version = 1;</code> 20206 */ 20207 boolean hasVersion(); 20208 /** 20209 * <code>optional uint32 version = 1;</code> 20210 */ 20211 int getVersion(); 20212 20213 // optional string owner = 2; 20214 /** 20215 * <code>optional string owner = 2;</code> 20216 */ 20217 boolean hasOwner(); 20218 /** 20219 * <code>optional string owner = 2;</code> 20220 */ 20221 java.lang.String getOwner(); 20222 /** 20223 * <code>optional string owner = 2;</code> 20224 */ 20225 com.google.protobuf.ByteString 20226 getOwnerBytes(); 20227 20228 // optional string renewer = 3; 20229 /** 20230 * <code>optional string renewer = 3;</code> 20231 */ 20232 boolean hasRenewer(); 20233 /** 20234 * <code>optional string renewer = 3;</code> 20235 */ 20236 java.lang.String getRenewer(); 20237 /** 20238 * <code>optional string renewer = 3;</code> 20239 */ 20240 com.google.protobuf.ByteString 20241 getRenewerBytes(); 20242 20243 // optional string realUser = 4; 20244 /** 20245 * <code>optional string realUser = 4;</code> 20246 */ 20247 boolean hasRealUser(); 20248 /** 20249 * <code>optional string realUser = 4;</code> 20250 */ 20251 java.lang.String getRealUser(); 20252 /** 20253 * <code>optional string realUser = 4;</code> 20254 */ 20255 com.google.protobuf.ByteString 20256 getRealUserBytes(); 20257 20258 // optional uint64 issueDate = 5; 20259 /** 20260 * <code>optional uint64 issueDate = 5;</code> 20261 */ 20262 boolean hasIssueDate(); 20263 /** 20264 * <code>optional uint64 issueDate = 5;</code> 20265 */ 20266 long getIssueDate(); 20267 20268 // optional uint64 maxDate = 6; 20269 /** 20270 * <code>optional uint64 maxDate = 6;</code> 20271 */ 20272 boolean hasMaxDate(); 20273 /** 20274 * <code>optional uint64 maxDate = 6;</code> 20275 */ 20276 long getMaxDate(); 20277 20278 // optional uint32 sequenceNumber = 7; 20279 /** 20280 * <code>optional uint32 sequenceNumber = 7;</code> 20281 */ 20282 boolean hasSequenceNumber(); 20283 /** 20284 * <code>optional uint32 sequenceNumber = 7;</code> 20285 */ 20286 int getSequenceNumber(); 20287 20288 // optional uint32 masterKeyId = 8; 20289 /** 20290 * <code>optional uint32 masterKeyId = 8;</code> 20291 */ 20292 boolean hasMasterKeyId(); 20293 /** 20294 * <code>optional uint32 masterKeyId = 8;</code> 20295 */ 20296 int getMasterKeyId(); 20297 20298 // optional uint64 expiryDate = 9; 20299 /** 20300 * <code>optional uint64 expiryDate = 9;</code> 20301 */ 20302 boolean hasExpiryDate(); 20303 /** 20304 * <code>optional uint64 expiryDate = 9;</code> 20305 */ 20306 long getExpiryDate(); 20307 } 20308 /** 20309 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.PersistToken} 20310 */ 20311 public static final class PersistToken extends 20312 com.google.protobuf.GeneratedMessage 20313 implements PersistTokenOrBuilder { 20314 // Use PersistToken.newBuilder() to construct. 20315 private PersistToken(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 20316 super(builder); 20317 this.unknownFields = builder.getUnknownFields(); 20318 } 20319 private PersistToken(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 20320 20321 private static final PersistToken defaultInstance; 20322 public static PersistToken getDefaultInstance() { 20323 return defaultInstance; 20324 } 20325 20326 public PersistToken getDefaultInstanceForType() { 20327 return defaultInstance; 20328 } 20329 20330 private final com.google.protobuf.UnknownFieldSet unknownFields; 20331 @java.lang.Override 20332 public final com.google.protobuf.UnknownFieldSet 20333 getUnknownFields() { 20334 return this.unknownFields; 20335 } 20336 private PersistToken( 20337 com.google.protobuf.CodedInputStream input, 20338 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20339 throws com.google.protobuf.InvalidProtocolBufferException { 20340 initFields(); 20341 int mutable_bitField0_ = 0; 20342 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 20343 com.google.protobuf.UnknownFieldSet.newBuilder(); 20344 try { 20345 boolean done = false; 20346 while (!done) { 20347 int tag = input.readTag(); 20348 switch (tag) { 20349 case 0: 20350 done = true; 20351 break; 20352 default: { 20353 if (!parseUnknownField(input, unknownFields, 20354 extensionRegistry, tag)) { 20355 done = true; 20356 } 20357 break; 20358 } 20359 case 8: { 20360 bitField0_ |= 0x00000001; 20361 version_ = input.readUInt32(); 20362 break; 20363 } 20364 case 18: { 20365 bitField0_ |= 0x00000002; 20366 owner_ = input.readBytes(); 20367 break; 20368 } 20369 case 26: { 20370 bitField0_ |= 0x00000004; 20371 renewer_ = input.readBytes(); 20372 break; 20373 } 20374 case 34: { 20375 bitField0_ |= 0x00000008; 20376 realUser_ = input.readBytes(); 20377 break; 20378 } 20379 case 40: { 20380 bitField0_ |= 0x00000010; 20381 issueDate_ = input.readUInt64(); 20382 break; 20383 } 20384 case 48: { 20385 bitField0_ |= 0x00000020; 20386 maxDate_ = input.readUInt64(); 20387 break; 20388 } 20389 case 56: { 20390 bitField0_ |= 0x00000040; 20391 sequenceNumber_ = input.readUInt32(); 20392 break; 20393 } 20394 case 64: { 20395 bitField0_ |= 0x00000080; 20396 masterKeyId_ = input.readUInt32(); 20397 break; 20398 } 20399 case 72: { 20400 bitField0_ |= 0x00000100; 20401 expiryDate_ = input.readUInt64(); 20402 break; 20403 } 20404 } 20405 } 20406 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 20407 throw e.setUnfinishedMessage(this); 20408 } catch (java.io.IOException e) { 20409 throw new com.google.protobuf.InvalidProtocolBufferException( 20410 e.getMessage()).setUnfinishedMessage(this); 20411 } finally { 20412 this.unknownFields = unknownFields.build(); 20413 makeExtensionsImmutable(); 20414 } 20415 } 20416 public static final com.google.protobuf.Descriptors.Descriptor 20417 getDescriptor() { 20418 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 20419 } 20420 20421 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 20422 internalGetFieldAccessorTable() { 20423 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable 20424 .ensureFieldAccessorsInitialized( 20425 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.Builder.class); 20426 } 20427 20428 public static com.google.protobuf.Parser<PersistToken> PARSER = 20429 new com.google.protobuf.AbstractParser<PersistToken>() { 20430 public PersistToken parsePartialFrom( 20431 com.google.protobuf.CodedInputStream input, 20432 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20433 throws com.google.protobuf.InvalidProtocolBufferException { 20434 return new PersistToken(input, extensionRegistry); 20435 } 20436 }; 20437 20438 @java.lang.Override 20439 public com.google.protobuf.Parser<PersistToken> getParserForType() { 20440 return PARSER; 20441 } 20442 20443 private int bitField0_; 20444 // optional uint32 version = 1; 20445 public static final int VERSION_FIELD_NUMBER = 1; 20446 private int version_; 20447 /** 20448 * <code>optional uint32 version = 1;</code> 20449 */ 20450 public boolean hasVersion() { 20451 return ((bitField0_ & 0x00000001) == 0x00000001); 20452 } 20453 /** 20454 * <code>optional uint32 version = 1;</code> 20455 */ 20456 public int getVersion() { 20457 return version_; 20458 } 20459 20460 // optional string owner = 2; 20461 public static final int OWNER_FIELD_NUMBER = 2; 20462 private java.lang.Object owner_; 20463 /** 20464 * <code>optional string owner = 2;</code> 20465 */ 20466 public boolean hasOwner() { 20467 return ((bitField0_ & 0x00000002) == 0x00000002); 20468 } 20469 /** 20470 * <code>optional string owner = 2;</code> 20471 */ 20472 public java.lang.String getOwner() { 20473 java.lang.Object ref = owner_; 20474 if (ref instanceof java.lang.String) { 20475 return (java.lang.String) ref; 20476 } else { 20477 com.google.protobuf.ByteString bs = 20478 (com.google.protobuf.ByteString) ref; 20479 java.lang.String s = bs.toStringUtf8(); 20480 if (bs.isValidUtf8()) { 20481 owner_ = s; 20482 } 20483 return s; 20484 } 20485 } 20486 /** 20487 * <code>optional string owner = 2;</code> 20488 */ 20489 public com.google.protobuf.ByteString 20490 getOwnerBytes() { 20491 java.lang.Object ref = owner_; 20492 if (ref instanceof java.lang.String) { 20493 com.google.protobuf.ByteString b = 20494 com.google.protobuf.ByteString.copyFromUtf8( 20495 (java.lang.String) ref); 20496 owner_ = b; 20497 return b; 20498 } else { 20499 return (com.google.protobuf.ByteString) ref; 20500 } 20501 } 20502 20503 // optional string renewer = 3; 20504 public static final int RENEWER_FIELD_NUMBER = 3; 20505 private java.lang.Object renewer_; 20506 /** 20507 * <code>optional string renewer = 3;</code> 20508 */ 20509 public boolean hasRenewer() { 20510 return ((bitField0_ & 0x00000004) == 0x00000004); 20511 } 20512 /** 20513 * <code>optional string renewer = 3;</code> 20514 */ 20515 public java.lang.String getRenewer() { 20516 java.lang.Object ref = renewer_; 20517 if (ref instanceof java.lang.String) { 20518 return (java.lang.String) ref; 20519 } else { 20520 com.google.protobuf.ByteString bs = 20521 (com.google.protobuf.ByteString) ref; 20522 java.lang.String s = bs.toStringUtf8(); 20523 if (bs.isValidUtf8()) { 20524 renewer_ = s; 20525 } 20526 return s; 20527 } 20528 } 20529 /** 20530 * <code>optional string renewer = 3;</code> 20531 */ 20532 public com.google.protobuf.ByteString 20533 getRenewerBytes() { 20534 java.lang.Object ref = renewer_; 20535 if (ref instanceof java.lang.String) { 20536 com.google.protobuf.ByteString b = 20537 com.google.protobuf.ByteString.copyFromUtf8( 20538 (java.lang.String) ref); 20539 renewer_ = b; 20540 return b; 20541 } else { 20542 return (com.google.protobuf.ByteString) ref; 20543 } 20544 } 20545 20546 // optional string realUser = 4; 20547 public static final int REALUSER_FIELD_NUMBER = 4; 20548 private java.lang.Object realUser_; 20549 /** 20550 * <code>optional string realUser = 4;</code> 20551 */ 20552 public boolean hasRealUser() { 20553 return ((bitField0_ & 0x00000008) == 0x00000008); 20554 } 20555 /** 20556 * <code>optional string realUser = 4;</code> 20557 */ 20558 public java.lang.String getRealUser() { 20559 java.lang.Object ref = realUser_; 20560 if (ref instanceof java.lang.String) { 20561 return (java.lang.String) ref; 20562 } else { 20563 com.google.protobuf.ByteString bs = 20564 (com.google.protobuf.ByteString) ref; 20565 java.lang.String s = bs.toStringUtf8(); 20566 if (bs.isValidUtf8()) { 20567 realUser_ = s; 20568 } 20569 return s; 20570 } 20571 } 20572 /** 20573 * <code>optional string realUser = 4;</code> 20574 */ 20575 public com.google.protobuf.ByteString 20576 getRealUserBytes() { 20577 java.lang.Object ref = realUser_; 20578 if (ref instanceof java.lang.String) { 20579 com.google.protobuf.ByteString b = 20580 com.google.protobuf.ByteString.copyFromUtf8( 20581 (java.lang.String) ref); 20582 realUser_ = b; 20583 return b; 20584 } else { 20585 return (com.google.protobuf.ByteString) ref; 20586 } 20587 } 20588 20589 // optional uint64 issueDate = 5; 20590 public static final int ISSUEDATE_FIELD_NUMBER = 5; 20591 private long issueDate_; 20592 /** 20593 * <code>optional uint64 issueDate = 5;</code> 20594 */ 20595 public boolean hasIssueDate() { 20596 return ((bitField0_ & 0x00000010) == 0x00000010); 20597 } 20598 /** 20599 * <code>optional uint64 issueDate = 5;</code> 20600 */ 20601 public long getIssueDate() { 20602 return issueDate_; 20603 } 20604 20605 // optional uint64 maxDate = 6; 20606 public static final int MAXDATE_FIELD_NUMBER = 6; 20607 private long maxDate_; 20608 /** 20609 * <code>optional uint64 maxDate = 6;</code> 20610 */ 20611 public boolean hasMaxDate() { 20612 return ((bitField0_ & 0x00000020) == 0x00000020); 20613 } 20614 /** 20615 * <code>optional uint64 maxDate = 6;</code> 20616 */ 20617 public long getMaxDate() { 20618 return maxDate_; 20619 } 20620 20621 // optional uint32 sequenceNumber = 7; 20622 public static final int SEQUENCENUMBER_FIELD_NUMBER = 7; 20623 private int sequenceNumber_; 20624 /** 20625 * <code>optional uint32 sequenceNumber = 7;</code> 20626 */ 20627 public boolean hasSequenceNumber() { 20628 return ((bitField0_ & 0x00000040) == 0x00000040); 20629 } 20630 /** 20631 * <code>optional uint32 sequenceNumber = 7;</code> 20632 */ 20633 public int getSequenceNumber() { 20634 return sequenceNumber_; 20635 } 20636 20637 // optional uint32 masterKeyId = 8; 20638 public static final int MASTERKEYID_FIELD_NUMBER = 8; 20639 private int masterKeyId_; 20640 /** 20641 * <code>optional uint32 masterKeyId = 8;</code> 20642 */ 20643 public boolean hasMasterKeyId() { 20644 return ((bitField0_ & 0x00000080) == 0x00000080); 20645 } 20646 /** 20647 * <code>optional uint32 masterKeyId = 8;</code> 20648 */ 20649 public int getMasterKeyId() { 20650 return masterKeyId_; 20651 } 20652 20653 // optional uint64 expiryDate = 9; 20654 public static final int EXPIRYDATE_FIELD_NUMBER = 9; 20655 private long expiryDate_; 20656 /** 20657 * <code>optional uint64 expiryDate = 9;</code> 20658 */ 20659 public boolean hasExpiryDate() { 20660 return ((bitField0_ & 0x00000100) == 0x00000100); 20661 } 20662 /** 20663 * <code>optional uint64 expiryDate = 9;</code> 20664 */ 20665 public long getExpiryDate() { 20666 return expiryDate_; 20667 } 20668 20669 private void initFields() { 20670 version_ = 0; 20671 owner_ = ""; 20672 renewer_ = ""; 20673 realUser_ = ""; 20674 issueDate_ = 0L; 20675 maxDate_ = 0L; 20676 sequenceNumber_ = 0; 20677 masterKeyId_ = 0; 20678 expiryDate_ = 0L; 20679 } 20680 private byte memoizedIsInitialized = -1; 20681 public final boolean isInitialized() { 20682 byte isInitialized = memoizedIsInitialized; 20683 if (isInitialized != -1) return isInitialized == 1; 20684 20685 memoizedIsInitialized = 1; 20686 return true; 20687 } 20688 20689 public void writeTo(com.google.protobuf.CodedOutputStream output) 20690 throws java.io.IOException { 20691 getSerializedSize(); 20692 if (((bitField0_ & 0x00000001) == 0x00000001)) { 20693 output.writeUInt32(1, version_); 20694 } 20695 if (((bitField0_ & 0x00000002) == 0x00000002)) { 20696 output.writeBytes(2, getOwnerBytes()); 20697 } 20698 if (((bitField0_ & 0x00000004) == 0x00000004)) { 20699 output.writeBytes(3, getRenewerBytes()); 20700 } 20701 if (((bitField0_ & 0x00000008) == 0x00000008)) { 20702 output.writeBytes(4, getRealUserBytes()); 20703 } 20704 if (((bitField0_ & 0x00000010) == 0x00000010)) { 20705 output.writeUInt64(5, issueDate_); 20706 } 20707 if (((bitField0_ & 0x00000020) == 0x00000020)) { 20708 output.writeUInt64(6, maxDate_); 20709 } 20710 if (((bitField0_ & 0x00000040) == 0x00000040)) { 20711 output.writeUInt32(7, sequenceNumber_); 20712 } 20713 if (((bitField0_ & 0x00000080) == 0x00000080)) { 20714 output.writeUInt32(8, masterKeyId_); 20715 } 20716 if (((bitField0_ & 0x00000100) == 0x00000100)) { 20717 output.writeUInt64(9, expiryDate_); 20718 } 20719 getUnknownFields().writeTo(output); 20720 } 20721 20722 private int memoizedSerializedSize = -1; 20723 public int getSerializedSize() { 20724 int size = memoizedSerializedSize; 20725 if (size != -1) return size; 20726 20727 size = 0; 20728 if (((bitField0_ & 0x00000001) == 0x00000001)) { 20729 size += com.google.protobuf.CodedOutputStream 20730 .computeUInt32Size(1, version_); 20731 } 20732 if (((bitField0_ & 0x00000002) == 0x00000002)) { 20733 size += com.google.protobuf.CodedOutputStream 20734 .computeBytesSize(2, getOwnerBytes()); 20735 } 20736 if (((bitField0_ & 0x00000004) == 0x00000004)) { 20737 size += com.google.protobuf.CodedOutputStream 20738 .computeBytesSize(3, getRenewerBytes()); 20739 } 20740 if (((bitField0_ & 0x00000008) == 0x00000008)) { 20741 size += com.google.protobuf.CodedOutputStream 20742 .computeBytesSize(4, getRealUserBytes()); 20743 } 20744 if (((bitField0_ & 0x00000010) == 0x00000010)) { 20745 size += com.google.protobuf.CodedOutputStream 20746 .computeUInt64Size(5, issueDate_); 20747 } 20748 if (((bitField0_ & 0x00000020) == 0x00000020)) { 20749 size += com.google.protobuf.CodedOutputStream 20750 .computeUInt64Size(6, maxDate_); 20751 } 20752 if (((bitField0_ & 0x00000040) == 0x00000040)) { 20753 size += com.google.protobuf.CodedOutputStream 20754 .computeUInt32Size(7, sequenceNumber_); 20755 } 20756 if (((bitField0_ & 0x00000080) == 0x00000080)) { 20757 size += com.google.protobuf.CodedOutputStream 20758 .computeUInt32Size(8, masterKeyId_); 20759 } 20760 if (((bitField0_ & 0x00000100) == 0x00000100)) { 20761 size += com.google.protobuf.CodedOutputStream 20762 .computeUInt64Size(9, expiryDate_); 20763 } 20764 size += getUnknownFields().getSerializedSize(); 20765 memoizedSerializedSize = size; 20766 return size; 20767 } 20768 20769 private static final long serialVersionUID = 0L; 20770 @java.lang.Override 20771 protected java.lang.Object writeReplace() 20772 throws java.io.ObjectStreamException { 20773 return super.writeReplace(); 20774 } 20775 20776 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20777 com.google.protobuf.ByteString data) 20778 throws com.google.protobuf.InvalidProtocolBufferException { 20779 return PARSER.parseFrom(data); 20780 } 20781 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20782 com.google.protobuf.ByteString data, 20783 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20784 throws com.google.protobuf.InvalidProtocolBufferException { 20785 return PARSER.parseFrom(data, extensionRegistry); 20786 } 20787 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom(byte[] data) 20788 throws com.google.protobuf.InvalidProtocolBufferException { 20789 return PARSER.parseFrom(data); 20790 } 20791 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20792 byte[] data, 20793 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20794 throws com.google.protobuf.InvalidProtocolBufferException { 20795 return PARSER.parseFrom(data, extensionRegistry); 20796 } 20797 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom(java.io.InputStream input) 20798 throws java.io.IOException { 20799 return PARSER.parseFrom(input); 20800 } 20801 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20802 java.io.InputStream input, 20803 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20804 throws java.io.IOException { 20805 return PARSER.parseFrom(input, extensionRegistry); 20806 } 20807 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseDelimitedFrom(java.io.InputStream input) 20808 throws java.io.IOException { 20809 return PARSER.parseDelimitedFrom(input); 20810 } 20811 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseDelimitedFrom( 20812 java.io.InputStream input, 20813 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20814 throws java.io.IOException { 20815 return PARSER.parseDelimitedFrom(input, extensionRegistry); 20816 } 20817 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20818 com.google.protobuf.CodedInputStream input) 20819 throws java.io.IOException { 20820 return PARSER.parseFrom(input); 20821 } 20822 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parseFrom( 20823 com.google.protobuf.CodedInputStream input, 20824 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 20825 throws java.io.IOException { 20826 return PARSER.parseFrom(input, extensionRegistry); 20827 } 20828 20829 public static Builder newBuilder() { return Builder.create(); } 20830 public Builder newBuilderForType() { return newBuilder(); } 20831 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken prototype) { 20832 return newBuilder().mergeFrom(prototype); 20833 } 20834 public Builder toBuilder() { return newBuilder(this); } 20835 20836 @java.lang.Override 20837 protected Builder newBuilderForType( 20838 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 20839 Builder builder = new Builder(parent); 20840 return builder; 20841 } 20842 /** 20843 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection.PersistToken} 20844 */ 20845 public static final class Builder extends 20846 com.google.protobuf.GeneratedMessage.Builder<Builder> 20847 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistTokenOrBuilder { 20848 public static final com.google.protobuf.Descriptors.Descriptor 20849 getDescriptor() { 20850 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 20851 } 20852 20853 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 20854 internalGetFieldAccessorTable() { 20855 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable 20856 .ensureFieldAccessorsInitialized( 20857 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.Builder.class); 20858 } 20859 20860 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.newBuilder() 20861 private Builder() { 20862 maybeForceBuilderInitialization(); 20863 } 20864 20865 private Builder( 20866 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 20867 super(parent); 20868 maybeForceBuilderInitialization(); 20869 } 20870 private void maybeForceBuilderInitialization() { 20871 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 20872 } 20873 } 20874 private static Builder create() { 20875 return new Builder(); 20876 } 20877 20878 public Builder clear() { 20879 super.clear(); 20880 version_ = 0; 20881 bitField0_ = (bitField0_ & ~0x00000001); 20882 owner_ = ""; 20883 bitField0_ = (bitField0_ & ~0x00000002); 20884 renewer_ = ""; 20885 bitField0_ = (bitField0_ & ~0x00000004); 20886 realUser_ = ""; 20887 bitField0_ = (bitField0_ & ~0x00000008); 20888 issueDate_ = 0L; 20889 bitField0_ = (bitField0_ & ~0x00000010); 20890 maxDate_ = 0L; 20891 bitField0_ = (bitField0_ & ~0x00000020); 20892 sequenceNumber_ = 0; 20893 bitField0_ = (bitField0_ & ~0x00000040); 20894 masterKeyId_ = 0; 20895 bitField0_ = (bitField0_ & ~0x00000080); 20896 expiryDate_ = 0L; 20897 bitField0_ = (bitField0_ & ~0x00000100); 20898 return this; 20899 } 20900 20901 public Builder clone() { 20902 return create().mergeFrom(buildPartial()); 20903 } 20904 20905 public com.google.protobuf.Descriptors.Descriptor 20906 getDescriptorForType() { 20907 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 20908 } 20909 20910 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken getDefaultInstanceForType() { 20911 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.getDefaultInstance(); 20912 } 20913 20914 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken build() { 20915 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken result = buildPartial(); 20916 if (!result.isInitialized()) { 20917 throw newUninitializedMessageException(result); 20918 } 20919 return result; 20920 } 20921 20922 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken buildPartial() { 20923 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken(this); 20924 int from_bitField0_ = bitField0_; 20925 int to_bitField0_ = 0; 20926 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 20927 to_bitField0_ |= 0x00000001; 20928 } 20929 result.version_ = version_; 20930 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 20931 to_bitField0_ |= 0x00000002; 20932 } 20933 result.owner_ = owner_; 20934 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 20935 to_bitField0_ |= 0x00000004; 20936 } 20937 result.renewer_ = renewer_; 20938 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 20939 to_bitField0_ |= 0x00000008; 20940 } 20941 result.realUser_ = realUser_; 20942 if (((from_bitField0_ & 0x00000010) == 0x00000010)) { 20943 to_bitField0_ |= 0x00000010; 20944 } 20945 result.issueDate_ = issueDate_; 20946 if (((from_bitField0_ & 0x00000020) == 0x00000020)) { 20947 to_bitField0_ |= 0x00000020; 20948 } 20949 result.maxDate_ = maxDate_; 20950 if (((from_bitField0_ & 0x00000040) == 0x00000040)) { 20951 to_bitField0_ |= 0x00000040; 20952 } 20953 result.sequenceNumber_ = sequenceNumber_; 20954 if (((from_bitField0_ & 0x00000080) == 0x00000080)) { 20955 to_bitField0_ |= 0x00000080; 20956 } 20957 result.masterKeyId_ = masterKeyId_; 20958 if (((from_bitField0_ & 0x00000100) == 0x00000100)) { 20959 to_bitField0_ |= 0x00000100; 20960 } 20961 result.expiryDate_ = expiryDate_; 20962 result.bitField0_ = to_bitField0_; 20963 onBuilt(); 20964 return result; 20965 } 20966 20967 public Builder mergeFrom(com.google.protobuf.Message other) { 20968 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken) { 20969 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken)other); 20970 } else { 20971 super.mergeFrom(other); 20972 return this; 20973 } 20974 } 20975 20976 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken other) { 20977 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken.getDefaultInstance()) return this; 20978 if (other.hasVersion()) { 20979 setVersion(other.getVersion()); 20980 } 20981 if (other.hasOwner()) { 20982 bitField0_ |= 0x00000002; 20983 owner_ = other.owner_; 20984 onChanged(); 20985 } 20986 if (other.hasRenewer()) { 20987 bitField0_ |= 0x00000004; 20988 renewer_ = other.renewer_; 20989 onChanged(); 20990 } 20991 if (other.hasRealUser()) { 20992 bitField0_ |= 0x00000008; 20993 realUser_ = other.realUser_; 20994 onChanged(); 20995 } 20996 if (other.hasIssueDate()) { 20997 setIssueDate(other.getIssueDate()); 20998 } 20999 if (other.hasMaxDate()) { 21000 setMaxDate(other.getMaxDate()); 21001 } 21002 if (other.hasSequenceNumber()) { 21003 setSequenceNumber(other.getSequenceNumber()); 21004 } 21005 if (other.hasMasterKeyId()) { 21006 setMasterKeyId(other.getMasterKeyId()); 21007 } 21008 if (other.hasExpiryDate()) { 21009 setExpiryDate(other.getExpiryDate()); 21010 } 21011 this.mergeUnknownFields(other.getUnknownFields()); 21012 return this; 21013 } 21014 21015 public final boolean isInitialized() { 21016 return true; 21017 } 21018 21019 public Builder mergeFrom( 21020 com.google.protobuf.CodedInputStream input, 21021 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21022 throws java.io.IOException { 21023 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken parsedMessage = null; 21024 try { 21025 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 21026 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 21027 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.PersistToken) e.getUnfinishedMessage(); 21028 throw e; 21029 } finally { 21030 if (parsedMessage != null) { 21031 mergeFrom(parsedMessage); 21032 } 21033 } 21034 return this; 21035 } 21036 private int bitField0_; 21037 21038 // optional uint32 version = 1; 21039 private int version_ ; 21040 /** 21041 * <code>optional uint32 version = 1;</code> 21042 */ 21043 public boolean hasVersion() { 21044 return ((bitField0_ & 0x00000001) == 0x00000001); 21045 } 21046 /** 21047 * <code>optional uint32 version = 1;</code> 21048 */ 21049 public int getVersion() { 21050 return version_; 21051 } 21052 /** 21053 * <code>optional uint32 version = 1;</code> 21054 */ 21055 public Builder setVersion(int value) { 21056 bitField0_ |= 0x00000001; 21057 version_ = value; 21058 onChanged(); 21059 return this; 21060 } 21061 /** 21062 * <code>optional uint32 version = 1;</code> 21063 */ 21064 public Builder clearVersion() { 21065 bitField0_ = (bitField0_ & ~0x00000001); 21066 version_ = 0; 21067 onChanged(); 21068 return this; 21069 } 21070 21071 // optional string owner = 2; 21072 private java.lang.Object owner_ = ""; 21073 /** 21074 * <code>optional string owner = 2;</code> 21075 */ 21076 public boolean hasOwner() { 21077 return ((bitField0_ & 0x00000002) == 0x00000002); 21078 } 21079 /** 21080 * <code>optional string owner = 2;</code> 21081 */ 21082 public java.lang.String getOwner() { 21083 java.lang.Object ref = owner_; 21084 if (!(ref instanceof java.lang.String)) { 21085 java.lang.String s = ((com.google.protobuf.ByteString) ref) 21086 .toStringUtf8(); 21087 owner_ = s; 21088 return s; 21089 } else { 21090 return (java.lang.String) ref; 21091 } 21092 } 21093 /** 21094 * <code>optional string owner = 2;</code> 21095 */ 21096 public com.google.protobuf.ByteString 21097 getOwnerBytes() { 21098 java.lang.Object ref = owner_; 21099 if (ref instanceof String) { 21100 com.google.protobuf.ByteString b = 21101 com.google.protobuf.ByteString.copyFromUtf8( 21102 (java.lang.String) ref); 21103 owner_ = b; 21104 return b; 21105 } else { 21106 return (com.google.protobuf.ByteString) ref; 21107 } 21108 } 21109 /** 21110 * <code>optional string owner = 2;</code> 21111 */ 21112 public Builder setOwner( 21113 java.lang.String value) { 21114 if (value == null) { 21115 throw new NullPointerException(); 21116 } 21117 bitField0_ |= 0x00000002; 21118 owner_ = value; 21119 onChanged(); 21120 return this; 21121 } 21122 /** 21123 * <code>optional string owner = 2;</code> 21124 */ 21125 public Builder clearOwner() { 21126 bitField0_ = (bitField0_ & ~0x00000002); 21127 owner_ = getDefaultInstance().getOwner(); 21128 onChanged(); 21129 return this; 21130 } 21131 /** 21132 * <code>optional string owner = 2;</code> 21133 */ 21134 public Builder setOwnerBytes( 21135 com.google.protobuf.ByteString value) { 21136 if (value == null) { 21137 throw new NullPointerException(); 21138 } 21139 bitField0_ |= 0x00000002; 21140 owner_ = value; 21141 onChanged(); 21142 return this; 21143 } 21144 21145 // optional string renewer = 3; 21146 private java.lang.Object renewer_ = ""; 21147 /** 21148 * <code>optional string renewer = 3;</code> 21149 */ 21150 public boolean hasRenewer() { 21151 return ((bitField0_ & 0x00000004) == 0x00000004); 21152 } 21153 /** 21154 * <code>optional string renewer = 3;</code> 21155 */ 21156 public java.lang.String getRenewer() { 21157 java.lang.Object ref = renewer_; 21158 if (!(ref instanceof java.lang.String)) { 21159 java.lang.String s = ((com.google.protobuf.ByteString) ref) 21160 .toStringUtf8(); 21161 renewer_ = s; 21162 return s; 21163 } else { 21164 return (java.lang.String) ref; 21165 } 21166 } 21167 /** 21168 * <code>optional string renewer = 3;</code> 21169 */ 21170 public com.google.protobuf.ByteString 21171 getRenewerBytes() { 21172 java.lang.Object ref = renewer_; 21173 if (ref instanceof String) { 21174 com.google.protobuf.ByteString b = 21175 com.google.protobuf.ByteString.copyFromUtf8( 21176 (java.lang.String) ref); 21177 renewer_ = b; 21178 return b; 21179 } else { 21180 return (com.google.protobuf.ByteString) ref; 21181 } 21182 } 21183 /** 21184 * <code>optional string renewer = 3;</code> 21185 */ 21186 public Builder setRenewer( 21187 java.lang.String value) { 21188 if (value == null) { 21189 throw new NullPointerException(); 21190 } 21191 bitField0_ |= 0x00000004; 21192 renewer_ = value; 21193 onChanged(); 21194 return this; 21195 } 21196 /** 21197 * <code>optional string renewer = 3;</code> 21198 */ 21199 public Builder clearRenewer() { 21200 bitField0_ = (bitField0_ & ~0x00000004); 21201 renewer_ = getDefaultInstance().getRenewer(); 21202 onChanged(); 21203 return this; 21204 } 21205 /** 21206 * <code>optional string renewer = 3;</code> 21207 */ 21208 public Builder setRenewerBytes( 21209 com.google.protobuf.ByteString value) { 21210 if (value == null) { 21211 throw new NullPointerException(); 21212 } 21213 bitField0_ |= 0x00000004; 21214 renewer_ = value; 21215 onChanged(); 21216 return this; 21217 } 21218 21219 // optional string realUser = 4; 21220 private java.lang.Object realUser_ = ""; 21221 /** 21222 * <code>optional string realUser = 4;</code> 21223 */ 21224 public boolean hasRealUser() { 21225 return ((bitField0_ & 0x00000008) == 0x00000008); 21226 } 21227 /** 21228 * <code>optional string realUser = 4;</code> 21229 */ 21230 public java.lang.String getRealUser() { 21231 java.lang.Object ref = realUser_; 21232 if (!(ref instanceof java.lang.String)) { 21233 java.lang.String s = ((com.google.protobuf.ByteString) ref) 21234 .toStringUtf8(); 21235 realUser_ = s; 21236 return s; 21237 } else { 21238 return (java.lang.String) ref; 21239 } 21240 } 21241 /** 21242 * <code>optional string realUser = 4;</code> 21243 */ 21244 public com.google.protobuf.ByteString 21245 getRealUserBytes() { 21246 java.lang.Object ref = realUser_; 21247 if (ref instanceof String) { 21248 com.google.protobuf.ByteString b = 21249 com.google.protobuf.ByteString.copyFromUtf8( 21250 (java.lang.String) ref); 21251 realUser_ = b; 21252 return b; 21253 } else { 21254 return (com.google.protobuf.ByteString) ref; 21255 } 21256 } 21257 /** 21258 * <code>optional string realUser = 4;</code> 21259 */ 21260 public Builder setRealUser( 21261 java.lang.String value) { 21262 if (value == null) { 21263 throw new NullPointerException(); 21264 } 21265 bitField0_ |= 0x00000008; 21266 realUser_ = value; 21267 onChanged(); 21268 return this; 21269 } 21270 /** 21271 * <code>optional string realUser = 4;</code> 21272 */ 21273 public Builder clearRealUser() { 21274 bitField0_ = (bitField0_ & ~0x00000008); 21275 realUser_ = getDefaultInstance().getRealUser(); 21276 onChanged(); 21277 return this; 21278 } 21279 /** 21280 * <code>optional string realUser = 4;</code> 21281 */ 21282 public Builder setRealUserBytes( 21283 com.google.protobuf.ByteString value) { 21284 if (value == null) { 21285 throw new NullPointerException(); 21286 } 21287 bitField0_ |= 0x00000008; 21288 realUser_ = value; 21289 onChanged(); 21290 return this; 21291 } 21292 21293 // optional uint64 issueDate = 5; 21294 private long issueDate_ ; 21295 /** 21296 * <code>optional uint64 issueDate = 5;</code> 21297 */ 21298 public boolean hasIssueDate() { 21299 return ((bitField0_ & 0x00000010) == 0x00000010); 21300 } 21301 /** 21302 * <code>optional uint64 issueDate = 5;</code> 21303 */ 21304 public long getIssueDate() { 21305 return issueDate_; 21306 } 21307 /** 21308 * <code>optional uint64 issueDate = 5;</code> 21309 */ 21310 public Builder setIssueDate(long value) { 21311 bitField0_ |= 0x00000010; 21312 issueDate_ = value; 21313 onChanged(); 21314 return this; 21315 } 21316 /** 21317 * <code>optional uint64 issueDate = 5;</code> 21318 */ 21319 public Builder clearIssueDate() { 21320 bitField0_ = (bitField0_ & ~0x00000010); 21321 issueDate_ = 0L; 21322 onChanged(); 21323 return this; 21324 } 21325 21326 // optional uint64 maxDate = 6; 21327 private long maxDate_ ; 21328 /** 21329 * <code>optional uint64 maxDate = 6;</code> 21330 */ 21331 public boolean hasMaxDate() { 21332 return ((bitField0_ & 0x00000020) == 0x00000020); 21333 } 21334 /** 21335 * <code>optional uint64 maxDate = 6;</code> 21336 */ 21337 public long getMaxDate() { 21338 return maxDate_; 21339 } 21340 /** 21341 * <code>optional uint64 maxDate = 6;</code> 21342 */ 21343 public Builder setMaxDate(long value) { 21344 bitField0_ |= 0x00000020; 21345 maxDate_ = value; 21346 onChanged(); 21347 return this; 21348 } 21349 /** 21350 * <code>optional uint64 maxDate = 6;</code> 21351 */ 21352 public Builder clearMaxDate() { 21353 bitField0_ = (bitField0_ & ~0x00000020); 21354 maxDate_ = 0L; 21355 onChanged(); 21356 return this; 21357 } 21358 21359 // optional uint32 sequenceNumber = 7; 21360 private int sequenceNumber_ ; 21361 /** 21362 * <code>optional uint32 sequenceNumber = 7;</code> 21363 */ 21364 public boolean hasSequenceNumber() { 21365 return ((bitField0_ & 0x00000040) == 0x00000040); 21366 } 21367 /** 21368 * <code>optional uint32 sequenceNumber = 7;</code> 21369 */ 21370 public int getSequenceNumber() { 21371 return sequenceNumber_; 21372 } 21373 /** 21374 * <code>optional uint32 sequenceNumber = 7;</code> 21375 */ 21376 public Builder setSequenceNumber(int value) { 21377 bitField0_ |= 0x00000040; 21378 sequenceNumber_ = value; 21379 onChanged(); 21380 return this; 21381 } 21382 /** 21383 * <code>optional uint32 sequenceNumber = 7;</code> 21384 */ 21385 public Builder clearSequenceNumber() { 21386 bitField0_ = (bitField0_ & ~0x00000040); 21387 sequenceNumber_ = 0; 21388 onChanged(); 21389 return this; 21390 } 21391 21392 // optional uint32 masterKeyId = 8; 21393 private int masterKeyId_ ; 21394 /** 21395 * <code>optional uint32 masterKeyId = 8;</code> 21396 */ 21397 public boolean hasMasterKeyId() { 21398 return ((bitField0_ & 0x00000080) == 0x00000080); 21399 } 21400 /** 21401 * <code>optional uint32 masterKeyId = 8;</code> 21402 */ 21403 public int getMasterKeyId() { 21404 return masterKeyId_; 21405 } 21406 /** 21407 * <code>optional uint32 masterKeyId = 8;</code> 21408 */ 21409 public Builder setMasterKeyId(int value) { 21410 bitField0_ |= 0x00000080; 21411 masterKeyId_ = value; 21412 onChanged(); 21413 return this; 21414 } 21415 /** 21416 * <code>optional uint32 masterKeyId = 8;</code> 21417 */ 21418 public Builder clearMasterKeyId() { 21419 bitField0_ = (bitField0_ & ~0x00000080); 21420 masterKeyId_ = 0; 21421 onChanged(); 21422 return this; 21423 } 21424 21425 // optional uint64 expiryDate = 9; 21426 private long expiryDate_ ; 21427 /** 21428 * <code>optional uint64 expiryDate = 9;</code> 21429 */ 21430 public boolean hasExpiryDate() { 21431 return ((bitField0_ & 0x00000100) == 0x00000100); 21432 } 21433 /** 21434 * <code>optional uint64 expiryDate = 9;</code> 21435 */ 21436 public long getExpiryDate() { 21437 return expiryDate_; 21438 } 21439 /** 21440 * <code>optional uint64 expiryDate = 9;</code> 21441 */ 21442 public Builder setExpiryDate(long value) { 21443 bitField0_ |= 0x00000100; 21444 expiryDate_ = value; 21445 onChanged(); 21446 return this; 21447 } 21448 /** 21449 * <code>optional uint64 expiryDate = 9;</code> 21450 */ 21451 public Builder clearExpiryDate() { 21452 bitField0_ = (bitField0_ & ~0x00000100); 21453 expiryDate_ = 0L; 21454 onChanged(); 21455 return this; 21456 } 21457 21458 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection.PersistToken) 21459 } 21460 21461 static { 21462 defaultInstance = new PersistToken(true); 21463 defaultInstance.initFields(); 21464 } 21465 21466 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection.PersistToken) 21467 } 21468 21469 private int bitField0_; 21470 // optional uint32 currentId = 1; 21471 public static final int CURRENTID_FIELD_NUMBER = 1; 21472 private int currentId_; 21473 /** 21474 * <code>optional uint32 currentId = 1;</code> 21475 */ 21476 public boolean hasCurrentId() { 21477 return ((bitField0_ & 0x00000001) == 0x00000001); 21478 } 21479 /** 21480 * <code>optional uint32 currentId = 1;</code> 21481 */ 21482 public int getCurrentId() { 21483 return currentId_; 21484 } 21485 21486 // optional uint32 tokenSequenceNumber = 2; 21487 public static final int TOKENSEQUENCENUMBER_FIELD_NUMBER = 2; 21488 private int tokenSequenceNumber_; 21489 /** 21490 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21491 */ 21492 public boolean hasTokenSequenceNumber() { 21493 return ((bitField0_ & 0x00000002) == 0x00000002); 21494 } 21495 /** 21496 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21497 */ 21498 public int getTokenSequenceNumber() { 21499 return tokenSequenceNumber_; 21500 } 21501 21502 // optional uint32 numKeys = 3; 21503 public static final int NUMKEYS_FIELD_NUMBER = 3; 21504 private int numKeys_; 21505 /** 21506 * <code>optional uint32 numKeys = 3;</code> 21507 */ 21508 public boolean hasNumKeys() { 21509 return ((bitField0_ & 0x00000004) == 0x00000004); 21510 } 21511 /** 21512 * <code>optional uint32 numKeys = 3;</code> 21513 */ 21514 public int getNumKeys() { 21515 return numKeys_; 21516 } 21517 21518 // optional uint32 numTokens = 4; 21519 public static final int NUMTOKENS_FIELD_NUMBER = 4; 21520 private int numTokens_; 21521 /** 21522 * <code>optional uint32 numTokens = 4;</code> 21523 * 21524 * <pre> 21525 * repeated DelegationKey keys 21526 * repeated PersistToken tokens 21527 * </pre> 21528 */ 21529 public boolean hasNumTokens() { 21530 return ((bitField0_ & 0x00000008) == 0x00000008); 21531 } 21532 /** 21533 * <code>optional uint32 numTokens = 4;</code> 21534 * 21535 * <pre> 21536 * repeated DelegationKey keys 21537 * repeated PersistToken tokens 21538 * </pre> 21539 */ 21540 public int getNumTokens() { 21541 return numTokens_; 21542 } 21543 21544 private void initFields() { 21545 currentId_ = 0; 21546 tokenSequenceNumber_ = 0; 21547 numKeys_ = 0; 21548 numTokens_ = 0; 21549 } 21550 private byte memoizedIsInitialized = -1; 21551 public final boolean isInitialized() { 21552 byte isInitialized = memoizedIsInitialized; 21553 if (isInitialized != -1) return isInitialized == 1; 21554 21555 memoizedIsInitialized = 1; 21556 return true; 21557 } 21558 21559 public void writeTo(com.google.protobuf.CodedOutputStream output) 21560 throws java.io.IOException { 21561 getSerializedSize(); 21562 if (((bitField0_ & 0x00000001) == 0x00000001)) { 21563 output.writeUInt32(1, currentId_); 21564 } 21565 if (((bitField0_ & 0x00000002) == 0x00000002)) { 21566 output.writeUInt32(2, tokenSequenceNumber_); 21567 } 21568 if (((bitField0_ & 0x00000004) == 0x00000004)) { 21569 output.writeUInt32(3, numKeys_); 21570 } 21571 if (((bitField0_ & 0x00000008) == 0x00000008)) { 21572 output.writeUInt32(4, numTokens_); 21573 } 21574 getUnknownFields().writeTo(output); 21575 } 21576 21577 private int memoizedSerializedSize = -1; 21578 public int getSerializedSize() { 21579 int size = memoizedSerializedSize; 21580 if (size != -1) return size; 21581 21582 size = 0; 21583 if (((bitField0_ & 0x00000001) == 0x00000001)) { 21584 size += com.google.protobuf.CodedOutputStream 21585 .computeUInt32Size(1, currentId_); 21586 } 21587 if (((bitField0_ & 0x00000002) == 0x00000002)) { 21588 size += com.google.protobuf.CodedOutputStream 21589 .computeUInt32Size(2, tokenSequenceNumber_); 21590 } 21591 if (((bitField0_ & 0x00000004) == 0x00000004)) { 21592 size += com.google.protobuf.CodedOutputStream 21593 .computeUInt32Size(3, numKeys_); 21594 } 21595 if (((bitField0_ & 0x00000008) == 0x00000008)) { 21596 size += com.google.protobuf.CodedOutputStream 21597 .computeUInt32Size(4, numTokens_); 21598 } 21599 size += getUnknownFields().getSerializedSize(); 21600 memoizedSerializedSize = size; 21601 return size; 21602 } 21603 21604 private static final long serialVersionUID = 0L; 21605 @java.lang.Override 21606 protected java.lang.Object writeReplace() 21607 throws java.io.ObjectStreamException { 21608 return super.writeReplace(); 21609 } 21610 21611 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21612 com.google.protobuf.ByteString data) 21613 throws com.google.protobuf.InvalidProtocolBufferException { 21614 return PARSER.parseFrom(data); 21615 } 21616 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21617 com.google.protobuf.ByteString data, 21618 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21619 throws com.google.protobuf.InvalidProtocolBufferException { 21620 return PARSER.parseFrom(data, extensionRegistry); 21621 } 21622 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom(byte[] data) 21623 throws com.google.protobuf.InvalidProtocolBufferException { 21624 return PARSER.parseFrom(data); 21625 } 21626 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21627 byte[] data, 21628 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21629 throws com.google.protobuf.InvalidProtocolBufferException { 21630 return PARSER.parseFrom(data, extensionRegistry); 21631 } 21632 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom(java.io.InputStream input) 21633 throws java.io.IOException { 21634 return PARSER.parseFrom(input); 21635 } 21636 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21637 java.io.InputStream input, 21638 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21639 throws java.io.IOException { 21640 return PARSER.parseFrom(input, extensionRegistry); 21641 } 21642 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseDelimitedFrom(java.io.InputStream input) 21643 throws java.io.IOException { 21644 return PARSER.parseDelimitedFrom(input); 21645 } 21646 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseDelimitedFrom( 21647 java.io.InputStream input, 21648 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21649 throws java.io.IOException { 21650 return PARSER.parseDelimitedFrom(input, extensionRegistry); 21651 } 21652 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21653 com.google.protobuf.CodedInputStream input) 21654 throws java.io.IOException { 21655 return PARSER.parseFrom(input); 21656 } 21657 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parseFrom( 21658 com.google.protobuf.CodedInputStream input, 21659 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21660 throws java.io.IOException { 21661 return PARSER.parseFrom(input, extensionRegistry); 21662 } 21663 21664 public static Builder newBuilder() { return Builder.create(); } 21665 public Builder newBuilderForType() { return newBuilder(); } 21666 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection prototype) { 21667 return newBuilder().mergeFrom(prototype); 21668 } 21669 public Builder toBuilder() { return newBuilder(this); } 21670 21671 @java.lang.Override 21672 protected Builder newBuilderForType( 21673 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 21674 Builder builder = new Builder(parent); 21675 return builder; 21676 } 21677 /** 21678 * Protobuf type {@code hadoop.hdfs.fsimage.SecretManagerSection} 21679 */ 21680 public static final class Builder extends 21681 com.google.protobuf.GeneratedMessage.Builder<Builder> 21682 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSectionOrBuilder { 21683 public static final com.google.protobuf.Descriptors.Descriptor 21684 getDescriptor() { 21685 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 21686 } 21687 21688 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 21689 internalGetFieldAccessorTable() { 21690 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable 21691 .ensureFieldAccessorsInitialized( 21692 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.Builder.class); 21693 } 21694 21695 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.newBuilder() 21696 private Builder() { 21697 maybeForceBuilderInitialization(); 21698 } 21699 21700 private Builder( 21701 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 21702 super(parent); 21703 maybeForceBuilderInitialization(); 21704 } 21705 private void maybeForceBuilderInitialization() { 21706 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 21707 } 21708 } 21709 private static Builder create() { 21710 return new Builder(); 21711 } 21712 21713 public Builder clear() { 21714 super.clear(); 21715 currentId_ = 0; 21716 bitField0_ = (bitField0_ & ~0x00000001); 21717 tokenSequenceNumber_ = 0; 21718 bitField0_ = (bitField0_ & ~0x00000002); 21719 numKeys_ = 0; 21720 bitField0_ = (bitField0_ & ~0x00000004); 21721 numTokens_ = 0; 21722 bitField0_ = (bitField0_ & ~0x00000008); 21723 return this; 21724 } 21725 21726 public Builder clone() { 21727 return create().mergeFrom(buildPartial()); 21728 } 21729 21730 public com.google.protobuf.Descriptors.Descriptor 21731 getDescriptorForType() { 21732 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 21733 } 21734 21735 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection getDefaultInstanceForType() { 21736 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.getDefaultInstance(); 21737 } 21738 21739 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection build() { 21740 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection result = buildPartial(); 21741 if (!result.isInitialized()) { 21742 throw newUninitializedMessageException(result); 21743 } 21744 return result; 21745 } 21746 21747 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection buildPartial() { 21748 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection(this); 21749 int from_bitField0_ = bitField0_; 21750 int to_bitField0_ = 0; 21751 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 21752 to_bitField0_ |= 0x00000001; 21753 } 21754 result.currentId_ = currentId_; 21755 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 21756 to_bitField0_ |= 0x00000002; 21757 } 21758 result.tokenSequenceNumber_ = tokenSequenceNumber_; 21759 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 21760 to_bitField0_ |= 0x00000004; 21761 } 21762 result.numKeys_ = numKeys_; 21763 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 21764 to_bitField0_ |= 0x00000008; 21765 } 21766 result.numTokens_ = numTokens_; 21767 result.bitField0_ = to_bitField0_; 21768 onBuilt(); 21769 return result; 21770 } 21771 21772 public Builder mergeFrom(com.google.protobuf.Message other) { 21773 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection) { 21774 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection)other); 21775 } else { 21776 super.mergeFrom(other); 21777 return this; 21778 } 21779 } 21780 21781 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection other) { 21782 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection.getDefaultInstance()) return this; 21783 if (other.hasCurrentId()) { 21784 setCurrentId(other.getCurrentId()); 21785 } 21786 if (other.hasTokenSequenceNumber()) { 21787 setTokenSequenceNumber(other.getTokenSequenceNumber()); 21788 } 21789 if (other.hasNumKeys()) { 21790 setNumKeys(other.getNumKeys()); 21791 } 21792 if (other.hasNumTokens()) { 21793 setNumTokens(other.getNumTokens()); 21794 } 21795 this.mergeUnknownFields(other.getUnknownFields()); 21796 return this; 21797 } 21798 21799 public final boolean isInitialized() { 21800 return true; 21801 } 21802 21803 public Builder mergeFrom( 21804 com.google.protobuf.CodedInputStream input, 21805 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 21806 throws java.io.IOException { 21807 org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection parsedMessage = null; 21808 try { 21809 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 21810 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 21811 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.SecretManagerSection) e.getUnfinishedMessage(); 21812 throw e; 21813 } finally { 21814 if (parsedMessage != null) { 21815 mergeFrom(parsedMessage); 21816 } 21817 } 21818 return this; 21819 } 21820 private int bitField0_; 21821 21822 // optional uint32 currentId = 1; 21823 private int currentId_ ; 21824 /** 21825 * <code>optional uint32 currentId = 1;</code> 21826 */ 21827 public boolean hasCurrentId() { 21828 return ((bitField0_ & 0x00000001) == 0x00000001); 21829 } 21830 /** 21831 * <code>optional uint32 currentId = 1;</code> 21832 */ 21833 public int getCurrentId() { 21834 return currentId_; 21835 } 21836 /** 21837 * <code>optional uint32 currentId = 1;</code> 21838 */ 21839 public Builder setCurrentId(int value) { 21840 bitField0_ |= 0x00000001; 21841 currentId_ = value; 21842 onChanged(); 21843 return this; 21844 } 21845 /** 21846 * <code>optional uint32 currentId = 1;</code> 21847 */ 21848 public Builder clearCurrentId() { 21849 bitField0_ = (bitField0_ & ~0x00000001); 21850 currentId_ = 0; 21851 onChanged(); 21852 return this; 21853 } 21854 21855 // optional uint32 tokenSequenceNumber = 2; 21856 private int tokenSequenceNumber_ ; 21857 /** 21858 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21859 */ 21860 public boolean hasTokenSequenceNumber() { 21861 return ((bitField0_ & 0x00000002) == 0x00000002); 21862 } 21863 /** 21864 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21865 */ 21866 public int getTokenSequenceNumber() { 21867 return tokenSequenceNumber_; 21868 } 21869 /** 21870 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21871 */ 21872 public Builder setTokenSequenceNumber(int value) { 21873 bitField0_ |= 0x00000002; 21874 tokenSequenceNumber_ = value; 21875 onChanged(); 21876 return this; 21877 } 21878 /** 21879 * <code>optional uint32 tokenSequenceNumber = 2;</code> 21880 */ 21881 public Builder clearTokenSequenceNumber() { 21882 bitField0_ = (bitField0_ & ~0x00000002); 21883 tokenSequenceNumber_ = 0; 21884 onChanged(); 21885 return this; 21886 } 21887 21888 // optional uint32 numKeys = 3; 21889 private int numKeys_ ; 21890 /** 21891 * <code>optional uint32 numKeys = 3;</code> 21892 */ 21893 public boolean hasNumKeys() { 21894 return ((bitField0_ & 0x00000004) == 0x00000004); 21895 } 21896 /** 21897 * <code>optional uint32 numKeys = 3;</code> 21898 */ 21899 public int getNumKeys() { 21900 return numKeys_; 21901 } 21902 /** 21903 * <code>optional uint32 numKeys = 3;</code> 21904 */ 21905 public Builder setNumKeys(int value) { 21906 bitField0_ |= 0x00000004; 21907 numKeys_ = value; 21908 onChanged(); 21909 return this; 21910 } 21911 /** 21912 * <code>optional uint32 numKeys = 3;</code> 21913 */ 21914 public Builder clearNumKeys() { 21915 bitField0_ = (bitField0_ & ~0x00000004); 21916 numKeys_ = 0; 21917 onChanged(); 21918 return this; 21919 } 21920 21921 // optional uint32 numTokens = 4; 21922 private int numTokens_ ; 21923 /** 21924 * <code>optional uint32 numTokens = 4;</code> 21925 * 21926 * <pre> 21927 * repeated DelegationKey keys 21928 * repeated PersistToken tokens 21929 * </pre> 21930 */ 21931 public boolean hasNumTokens() { 21932 return ((bitField0_ & 0x00000008) == 0x00000008); 21933 } 21934 /** 21935 * <code>optional uint32 numTokens = 4;</code> 21936 * 21937 * <pre> 21938 * repeated DelegationKey keys 21939 * repeated PersistToken tokens 21940 * </pre> 21941 */ 21942 public int getNumTokens() { 21943 return numTokens_; 21944 } 21945 /** 21946 * <code>optional uint32 numTokens = 4;</code> 21947 * 21948 * <pre> 21949 * repeated DelegationKey keys 21950 * repeated PersistToken tokens 21951 * </pre> 21952 */ 21953 public Builder setNumTokens(int value) { 21954 bitField0_ |= 0x00000008; 21955 numTokens_ = value; 21956 onChanged(); 21957 return this; 21958 } 21959 /** 21960 * <code>optional uint32 numTokens = 4;</code> 21961 * 21962 * <pre> 21963 * repeated DelegationKey keys 21964 * repeated PersistToken tokens 21965 * </pre> 21966 */ 21967 public Builder clearNumTokens() { 21968 bitField0_ = (bitField0_ & ~0x00000008); 21969 numTokens_ = 0; 21970 onChanged(); 21971 return this; 21972 } 21973 21974 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.SecretManagerSection) 21975 } 21976 21977 static { 21978 defaultInstance = new SecretManagerSection(true); 21979 defaultInstance.initFields(); 21980 } 21981 21982 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.SecretManagerSection) 21983 } 21984 21985 public interface CacheManagerSectionOrBuilder 21986 extends com.google.protobuf.MessageOrBuilder { 21987 21988 // required uint64 nextDirectiveId = 1; 21989 /** 21990 * <code>required uint64 nextDirectiveId = 1;</code> 21991 */ 21992 boolean hasNextDirectiveId(); 21993 /** 21994 * <code>required uint64 nextDirectiveId = 1;</code> 21995 */ 21996 long getNextDirectiveId(); 21997 21998 // required uint32 numPools = 2; 21999 /** 22000 * <code>required uint32 numPools = 2;</code> 22001 */ 22002 boolean hasNumPools(); 22003 /** 22004 * <code>required uint32 numPools = 2;</code> 22005 */ 22006 int getNumPools(); 22007 22008 // required uint32 numDirectives = 3; 22009 /** 22010 * <code>required uint32 numDirectives = 3;</code> 22011 * 22012 * <pre> 22013 * repeated CachePoolInfoProto pools 22014 * repeated CacheDirectiveInfoProto directives 22015 * </pre> 22016 */ 22017 boolean hasNumDirectives(); 22018 /** 22019 * <code>required uint32 numDirectives = 3;</code> 22020 * 22021 * <pre> 22022 * repeated CachePoolInfoProto pools 22023 * repeated CacheDirectiveInfoProto directives 22024 * </pre> 22025 */ 22026 int getNumDirectives(); 22027 } 22028 /** 22029 * Protobuf type {@code hadoop.hdfs.fsimage.CacheManagerSection} 22030 */ 22031 public static final class CacheManagerSection extends 22032 com.google.protobuf.GeneratedMessage 22033 implements CacheManagerSectionOrBuilder { 22034 // Use CacheManagerSection.newBuilder() to construct. 22035 private CacheManagerSection(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 22036 super(builder); 22037 this.unknownFields = builder.getUnknownFields(); 22038 } 22039 private CacheManagerSection(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 22040 22041 private static final CacheManagerSection defaultInstance; 22042 public static CacheManagerSection getDefaultInstance() { 22043 return defaultInstance; 22044 } 22045 22046 public CacheManagerSection getDefaultInstanceForType() { 22047 return defaultInstance; 22048 } 22049 22050 private final com.google.protobuf.UnknownFieldSet unknownFields; 22051 @java.lang.Override 22052 public final com.google.protobuf.UnknownFieldSet 22053 getUnknownFields() { 22054 return this.unknownFields; 22055 } 22056 private CacheManagerSection( 22057 com.google.protobuf.CodedInputStream input, 22058 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22059 throws com.google.protobuf.InvalidProtocolBufferException { 22060 initFields(); 22061 int mutable_bitField0_ = 0; 22062 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 22063 com.google.protobuf.UnknownFieldSet.newBuilder(); 22064 try { 22065 boolean done = false; 22066 while (!done) { 22067 int tag = input.readTag(); 22068 switch (tag) { 22069 case 0: 22070 done = true; 22071 break; 22072 default: { 22073 if (!parseUnknownField(input, unknownFields, 22074 extensionRegistry, tag)) { 22075 done = true; 22076 } 22077 break; 22078 } 22079 case 8: { 22080 bitField0_ |= 0x00000001; 22081 nextDirectiveId_ = input.readUInt64(); 22082 break; 22083 } 22084 case 16: { 22085 bitField0_ |= 0x00000002; 22086 numPools_ = input.readUInt32(); 22087 break; 22088 } 22089 case 24: { 22090 bitField0_ |= 0x00000004; 22091 numDirectives_ = input.readUInt32(); 22092 break; 22093 } 22094 } 22095 } 22096 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 22097 throw e.setUnfinishedMessage(this); 22098 } catch (java.io.IOException e) { 22099 throw new com.google.protobuf.InvalidProtocolBufferException( 22100 e.getMessage()).setUnfinishedMessage(this); 22101 } finally { 22102 this.unknownFields = unknownFields.build(); 22103 makeExtensionsImmutable(); 22104 } 22105 } 22106 public static final com.google.protobuf.Descriptors.Descriptor 22107 getDescriptor() { 22108 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 22109 } 22110 22111 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 22112 internalGetFieldAccessorTable() { 22113 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable 22114 .ensureFieldAccessorsInitialized( 22115 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.Builder.class); 22116 } 22117 22118 public static com.google.protobuf.Parser<CacheManagerSection> PARSER = 22119 new com.google.protobuf.AbstractParser<CacheManagerSection>() { 22120 public CacheManagerSection parsePartialFrom( 22121 com.google.protobuf.CodedInputStream input, 22122 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22123 throws com.google.protobuf.InvalidProtocolBufferException { 22124 return new CacheManagerSection(input, extensionRegistry); 22125 } 22126 }; 22127 22128 @java.lang.Override 22129 public com.google.protobuf.Parser<CacheManagerSection> getParserForType() { 22130 return PARSER; 22131 } 22132 22133 private int bitField0_; 22134 // required uint64 nextDirectiveId = 1; 22135 public static final int NEXTDIRECTIVEID_FIELD_NUMBER = 1; 22136 private long nextDirectiveId_; 22137 /** 22138 * <code>required uint64 nextDirectiveId = 1;</code> 22139 */ 22140 public boolean hasNextDirectiveId() { 22141 return ((bitField0_ & 0x00000001) == 0x00000001); 22142 } 22143 /** 22144 * <code>required uint64 nextDirectiveId = 1;</code> 22145 */ 22146 public long getNextDirectiveId() { 22147 return nextDirectiveId_; 22148 } 22149 22150 // required uint32 numPools = 2; 22151 public static final int NUMPOOLS_FIELD_NUMBER = 2; 22152 private int numPools_; 22153 /** 22154 * <code>required uint32 numPools = 2;</code> 22155 */ 22156 public boolean hasNumPools() { 22157 return ((bitField0_ & 0x00000002) == 0x00000002); 22158 } 22159 /** 22160 * <code>required uint32 numPools = 2;</code> 22161 */ 22162 public int getNumPools() { 22163 return numPools_; 22164 } 22165 22166 // required uint32 numDirectives = 3; 22167 public static final int NUMDIRECTIVES_FIELD_NUMBER = 3; 22168 private int numDirectives_; 22169 /** 22170 * <code>required uint32 numDirectives = 3;</code> 22171 * 22172 * <pre> 22173 * repeated CachePoolInfoProto pools 22174 * repeated CacheDirectiveInfoProto directives 22175 * </pre> 22176 */ 22177 public boolean hasNumDirectives() { 22178 return ((bitField0_ & 0x00000004) == 0x00000004); 22179 } 22180 /** 22181 * <code>required uint32 numDirectives = 3;</code> 22182 * 22183 * <pre> 22184 * repeated CachePoolInfoProto pools 22185 * repeated CacheDirectiveInfoProto directives 22186 * </pre> 22187 */ 22188 public int getNumDirectives() { 22189 return numDirectives_; 22190 } 22191 22192 private void initFields() { 22193 nextDirectiveId_ = 0L; 22194 numPools_ = 0; 22195 numDirectives_ = 0; 22196 } 22197 private byte memoizedIsInitialized = -1; 22198 public final boolean isInitialized() { 22199 byte isInitialized = memoizedIsInitialized; 22200 if (isInitialized != -1) return isInitialized == 1; 22201 22202 if (!hasNextDirectiveId()) { 22203 memoizedIsInitialized = 0; 22204 return false; 22205 } 22206 if (!hasNumPools()) { 22207 memoizedIsInitialized = 0; 22208 return false; 22209 } 22210 if (!hasNumDirectives()) { 22211 memoizedIsInitialized = 0; 22212 return false; 22213 } 22214 memoizedIsInitialized = 1; 22215 return true; 22216 } 22217 22218 public void writeTo(com.google.protobuf.CodedOutputStream output) 22219 throws java.io.IOException { 22220 getSerializedSize(); 22221 if (((bitField0_ & 0x00000001) == 0x00000001)) { 22222 output.writeUInt64(1, nextDirectiveId_); 22223 } 22224 if (((bitField0_ & 0x00000002) == 0x00000002)) { 22225 output.writeUInt32(2, numPools_); 22226 } 22227 if (((bitField0_ & 0x00000004) == 0x00000004)) { 22228 output.writeUInt32(3, numDirectives_); 22229 } 22230 getUnknownFields().writeTo(output); 22231 } 22232 22233 private int memoizedSerializedSize = -1; 22234 public int getSerializedSize() { 22235 int size = memoizedSerializedSize; 22236 if (size != -1) return size; 22237 22238 size = 0; 22239 if (((bitField0_ & 0x00000001) == 0x00000001)) { 22240 size += com.google.protobuf.CodedOutputStream 22241 .computeUInt64Size(1, nextDirectiveId_); 22242 } 22243 if (((bitField0_ & 0x00000002) == 0x00000002)) { 22244 size += com.google.protobuf.CodedOutputStream 22245 .computeUInt32Size(2, numPools_); 22246 } 22247 if (((bitField0_ & 0x00000004) == 0x00000004)) { 22248 size += com.google.protobuf.CodedOutputStream 22249 .computeUInt32Size(3, numDirectives_); 22250 } 22251 size += getUnknownFields().getSerializedSize(); 22252 memoizedSerializedSize = size; 22253 return size; 22254 } 22255 22256 private static final long serialVersionUID = 0L; 22257 @java.lang.Override 22258 protected java.lang.Object writeReplace() 22259 throws java.io.ObjectStreamException { 22260 return super.writeReplace(); 22261 } 22262 22263 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22264 com.google.protobuf.ByteString data) 22265 throws com.google.protobuf.InvalidProtocolBufferException { 22266 return PARSER.parseFrom(data); 22267 } 22268 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22269 com.google.protobuf.ByteString data, 22270 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22271 throws com.google.protobuf.InvalidProtocolBufferException { 22272 return PARSER.parseFrom(data, extensionRegistry); 22273 } 22274 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom(byte[] data) 22275 throws com.google.protobuf.InvalidProtocolBufferException { 22276 return PARSER.parseFrom(data); 22277 } 22278 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22279 byte[] data, 22280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22281 throws com.google.protobuf.InvalidProtocolBufferException { 22282 return PARSER.parseFrom(data, extensionRegistry); 22283 } 22284 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom(java.io.InputStream input) 22285 throws java.io.IOException { 22286 return PARSER.parseFrom(input); 22287 } 22288 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22289 java.io.InputStream input, 22290 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22291 throws java.io.IOException { 22292 return PARSER.parseFrom(input, extensionRegistry); 22293 } 22294 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseDelimitedFrom(java.io.InputStream input) 22295 throws java.io.IOException { 22296 return PARSER.parseDelimitedFrom(input); 22297 } 22298 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseDelimitedFrom( 22299 java.io.InputStream input, 22300 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22301 throws java.io.IOException { 22302 return PARSER.parseDelimitedFrom(input, extensionRegistry); 22303 } 22304 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22305 com.google.protobuf.CodedInputStream input) 22306 throws java.io.IOException { 22307 return PARSER.parseFrom(input); 22308 } 22309 public static org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parseFrom( 22310 com.google.protobuf.CodedInputStream input, 22311 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22312 throws java.io.IOException { 22313 return PARSER.parseFrom(input, extensionRegistry); 22314 } 22315 22316 public static Builder newBuilder() { return Builder.create(); } 22317 public Builder newBuilderForType() { return newBuilder(); } 22318 public static Builder newBuilder(org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection prototype) { 22319 return newBuilder().mergeFrom(prototype); 22320 } 22321 public Builder toBuilder() { return newBuilder(this); } 22322 22323 @java.lang.Override 22324 protected Builder newBuilderForType( 22325 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 22326 Builder builder = new Builder(parent); 22327 return builder; 22328 } 22329 /** 22330 * Protobuf type {@code hadoop.hdfs.fsimage.CacheManagerSection} 22331 */ 22332 public static final class Builder extends 22333 com.google.protobuf.GeneratedMessage.Builder<Builder> 22334 implements org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSectionOrBuilder { 22335 public static final com.google.protobuf.Descriptors.Descriptor 22336 getDescriptor() { 22337 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 22338 } 22339 22340 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 22341 internalGetFieldAccessorTable() { 22342 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable 22343 .ensureFieldAccessorsInitialized( 22344 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.class, org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.Builder.class); 22345 } 22346 22347 // Construct using org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.newBuilder() 22348 private Builder() { 22349 maybeForceBuilderInitialization(); 22350 } 22351 22352 private Builder( 22353 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 22354 super(parent); 22355 maybeForceBuilderInitialization(); 22356 } 22357 private void maybeForceBuilderInitialization() { 22358 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 22359 } 22360 } 22361 private static Builder create() { 22362 return new Builder(); 22363 } 22364 22365 public Builder clear() { 22366 super.clear(); 22367 nextDirectiveId_ = 0L; 22368 bitField0_ = (bitField0_ & ~0x00000001); 22369 numPools_ = 0; 22370 bitField0_ = (bitField0_ & ~0x00000002); 22371 numDirectives_ = 0; 22372 bitField0_ = (bitField0_ & ~0x00000004); 22373 return this; 22374 } 22375 22376 public Builder clone() { 22377 return create().mergeFrom(buildPartial()); 22378 } 22379 22380 public com.google.protobuf.Descriptors.Descriptor 22381 getDescriptorForType() { 22382 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 22383 } 22384 22385 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection getDefaultInstanceForType() { 22386 return org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.getDefaultInstance(); 22387 } 22388 22389 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection build() { 22390 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection result = buildPartial(); 22391 if (!result.isInitialized()) { 22392 throw newUninitializedMessageException(result); 22393 } 22394 return result; 22395 } 22396 22397 public org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection buildPartial() { 22398 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection result = new org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection(this); 22399 int from_bitField0_ = bitField0_; 22400 int to_bitField0_ = 0; 22401 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 22402 to_bitField0_ |= 0x00000001; 22403 } 22404 result.nextDirectiveId_ = nextDirectiveId_; 22405 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 22406 to_bitField0_ |= 0x00000002; 22407 } 22408 result.numPools_ = numPools_; 22409 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 22410 to_bitField0_ |= 0x00000004; 22411 } 22412 result.numDirectives_ = numDirectives_; 22413 result.bitField0_ = to_bitField0_; 22414 onBuilt(); 22415 return result; 22416 } 22417 22418 public Builder mergeFrom(com.google.protobuf.Message other) { 22419 if (other instanceof org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection) { 22420 return mergeFrom((org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection)other); 22421 } else { 22422 super.mergeFrom(other); 22423 return this; 22424 } 22425 } 22426 22427 public Builder mergeFrom(org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection other) { 22428 if (other == org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection.getDefaultInstance()) return this; 22429 if (other.hasNextDirectiveId()) { 22430 setNextDirectiveId(other.getNextDirectiveId()); 22431 } 22432 if (other.hasNumPools()) { 22433 setNumPools(other.getNumPools()); 22434 } 22435 if (other.hasNumDirectives()) { 22436 setNumDirectives(other.getNumDirectives()); 22437 } 22438 this.mergeUnknownFields(other.getUnknownFields()); 22439 return this; 22440 } 22441 22442 public final boolean isInitialized() { 22443 if (!hasNextDirectiveId()) { 22444 22445 return false; 22446 } 22447 if (!hasNumPools()) { 22448 22449 return false; 22450 } 22451 if (!hasNumDirectives()) { 22452 22453 return false; 22454 } 22455 return true; 22456 } 22457 22458 public Builder mergeFrom( 22459 com.google.protobuf.CodedInputStream input, 22460 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 22461 throws java.io.IOException { 22462 org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection parsedMessage = null; 22463 try { 22464 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 22465 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 22466 parsedMessage = (org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection) e.getUnfinishedMessage(); 22467 throw e; 22468 } finally { 22469 if (parsedMessage != null) { 22470 mergeFrom(parsedMessage); 22471 } 22472 } 22473 return this; 22474 } 22475 private int bitField0_; 22476 22477 // required uint64 nextDirectiveId = 1; 22478 private long nextDirectiveId_ ; 22479 /** 22480 * <code>required uint64 nextDirectiveId = 1;</code> 22481 */ 22482 public boolean hasNextDirectiveId() { 22483 return ((bitField0_ & 0x00000001) == 0x00000001); 22484 } 22485 /** 22486 * <code>required uint64 nextDirectiveId = 1;</code> 22487 */ 22488 public long getNextDirectiveId() { 22489 return nextDirectiveId_; 22490 } 22491 /** 22492 * <code>required uint64 nextDirectiveId = 1;</code> 22493 */ 22494 public Builder setNextDirectiveId(long value) { 22495 bitField0_ |= 0x00000001; 22496 nextDirectiveId_ = value; 22497 onChanged(); 22498 return this; 22499 } 22500 /** 22501 * <code>required uint64 nextDirectiveId = 1;</code> 22502 */ 22503 public Builder clearNextDirectiveId() { 22504 bitField0_ = (bitField0_ & ~0x00000001); 22505 nextDirectiveId_ = 0L; 22506 onChanged(); 22507 return this; 22508 } 22509 22510 // required uint32 numPools = 2; 22511 private int numPools_ ; 22512 /** 22513 * <code>required uint32 numPools = 2;</code> 22514 */ 22515 public boolean hasNumPools() { 22516 return ((bitField0_ & 0x00000002) == 0x00000002); 22517 } 22518 /** 22519 * <code>required uint32 numPools = 2;</code> 22520 */ 22521 public int getNumPools() { 22522 return numPools_; 22523 } 22524 /** 22525 * <code>required uint32 numPools = 2;</code> 22526 */ 22527 public Builder setNumPools(int value) { 22528 bitField0_ |= 0x00000002; 22529 numPools_ = value; 22530 onChanged(); 22531 return this; 22532 } 22533 /** 22534 * <code>required uint32 numPools = 2;</code> 22535 */ 22536 public Builder clearNumPools() { 22537 bitField0_ = (bitField0_ & ~0x00000002); 22538 numPools_ = 0; 22539 onChanged(); 22540 return this; 22541 } 22542 22543 // required uint32 numDirectives = 3; 22544 private int numDirectives_ ; 22545 /** 22546 * <code>required uint32 numDirectives = 3;</code> 22547 * 22548 * <pre> 22549 * repeated CachePoolInfoProto pools 22550 * repeated CacheDirectiveInfoProto directives 22551 * </pre> 22552 */ 22553 public boolean hasNumDirectives() { 22554 return ((bitField0_ & 0x00000004) == 0x00000004); 22555 } 22556 /** 22557 * <code>required uint32 numDirectives = 3;</code> 22558 * 22559 * <pre> 22560 * repeated CachePoolInfoProto pools 22561 * repeated CacheDirectiveInfoProto directives 22562 * </pre> 22563 */ 22564 public int getNumDirectives() { 22565 return numDirectives_; 22566 } 22567 /** 22568 * <code>required uint32 numDirectives = 3;</code> 22569 * 22570 * <pre> 22571 * repeated CachePoolInfoProto pools 22572 * repeated CacheDirectiveInfoProto directives 22573 * </pre> 22574 */ 22575 public Builder setNumDirectives(int value) { 22576 bitField0_ |= 0x00000004; 22577 numDirectives_ = value; 22578 onChanged(); 22579 return this; 22580 } 22581 /** 22582 * <code>required uint32 numDirectives = 3;</code> 22583 * 22584 * <pre> 22585 * repeated CachePoolInfoProto pools 22586 * repeated CacheDirectiveInfoProto directives 22587 * </pre> 22588 */ 22589 public Builder clearNumDirectives() { 22590 bitField0_ = (bitField0_ & ~0x00000004); 22591 numDirectives_ = 0; 22592 onChanged(); 22593 return this; 22594 } 22595 22596 // @@protoc_insertion_point(builder_scope:hadoop.hdfs.fsimage.CacheManagerSection) 22597 } 22598 22599 static { 22600 defaultInstance = new CacheManagerSection(true); 22601 defaultInstance.initFields(); 22602 } 22603 22604 // @@protoc_insertion_point(class_scope:hadoop.hdfs.fsimage.CacheManagerSection) 22605 } 22606 22607 private static com.google.protobuf.Descriptors.Descriptor 22608 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor; 22609 private static 22610 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22611 internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable; 22612 private static com.google.protobuf.Descriptors.Descriptor 22613 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor; 22614 private static 22615 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22616 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable; 22617 private static com.google.protobuf.Descriptors.Descriptor 22618 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor; 22619 private static 22620 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22621 internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable; 22622 private static com.google.protobuf.Descriptors.Descriptor 22623 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor; 22624 private static 22625 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22626 internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable; 22627 private static com.google.protobuf.Descriptors.Descriptor 22628 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor; 22629 private static 22630 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22631 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable; 22632 private static com.google.protobuf.Descriptors.Descriptor 22633 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor; 22634 private static 22635 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22636 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable; 22637 private static com.google.protobuf.Descriptors.Descriptor 22638 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor; 22639 private static 22640 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22641 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_fieldAccessorTable; 22642 private static com.google.protobuf.Descriptors.Descriptor 22643 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor; 22644 private static 22645 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22646 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_fieldAccessorTable; 22647 private static com.google.protobuf.Descriptors.Descriptor 22648 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor; 22649 private static 22650 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22651 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable; 22652 private static com.google.protobuf.Descriptors.Descriptor 22653 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor; 22654 private static 22655 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22656 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable; 22657 private static com.google.protobuf.Descriptors.Descriptor 22658 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor; 22659 private static 22660 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22661 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable; 22662 private static com.google.protobuf.Descriptors.Descriptor 22663 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor; 22664 private static 22665 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22666 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable; 22667 private static com.google.protobuf.Descriptors.Descriptor 22668 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor; 22669 private static 22670 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22671 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable; 22672 private static com.google.protobuf.Descriptors.Descriptor 22673 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor; 22674 private static 22675 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22676 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable; 22677 private static com.google.protobuf.Descriptors.Descriptor 22678 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor; 22679 private static 22680 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22681 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable; 22682 private static com.google.protobuf.Descriptors.Descriptor 22683 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor; 22684 private static 22685 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22686 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable; 22687 private static com.google.protobuf.Descriptors.Descriptor 22688 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor; 22689 private static 22690 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22691 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable; 22692 private static com.google.protobuf.Descriptors.Descriptor 22693 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor; 22694 private static 22695 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22696 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable; 22697 private static com.google.protobuf.Descriptors.Descriptor 22698 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor; 22699 private static 22700 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22701 internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable; 22702 private static com.google.protobuf.Descriptors.Descriptor 22703 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor; 22704 private static 22705 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22706 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable; 22707 private static com.google.protobuf.Descriptors.Descriptor 22708 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor; 22709 private static 22710 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22711 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable; 22712 private static com.google.protobuf.Descriptors.Descriptor 22713 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor; 22714 private static 22715 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22716 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable; 22717 private static com.google.protobuf.Descriptors.Descriptor 22718 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor; 22719 private static 22720 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22721 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable; 22722 private static com.google.protobuf.Descriptors.Descriptor 22723 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor; 22724 private static 22725 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22726 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable; 22727 private static com.google.protobuf.Descriptors.Descriptor 22728 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor; 22729 private static 22730 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22731 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable; 22732 private static com.google.protobuf.Descriptors.Descriptor 22733 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor; 22734 private static 22735 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22736 internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable; 22737 private static com.google.protobuf.Descriptors.Descriptor 22738 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor; 22739 private static 22740 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22741 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable; 22742 private static com.google.protobuf.Descriptors.Descriptor 22743 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor; 22744 private static 22745 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22746 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable; 22747 private static com.google.protobuf.Descriptors.Descriptor 22748 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor; 22749 private static 22750 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22751 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable; 22752 private static com.google.protobuf.Descriptors.Descriptor 22753 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor; 22754 private static 22755 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22756 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable; 22757 private static com.google.protobuf.Descriptors.Descriptor 22758 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor; 22759 private static 22760 com.google.protobuf.GeneratedMessage.FieldAccessorTable 22761 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable; 22762 22763 public static com.google.protobuf.Descriptors.FileDescriptor 22764 getDescriptor() { 22765 return descriptor; 22766 } 22767 private static com.google.protobuf.Descriptors.FileDescriptor 22768 descriptor; 22769 static { 22770 java.lang.String[] descriptorData = { 22771 "\n\rfsimage.proto\022\023hadoop.hdfs.fsimage\032\nhd" + 22772 "fs.proto\032\tacl.proto\032\013xattr.proto\"\277\001\n\013Fil" + 22773 "eSummary\022\025\n\rondiskVersion\030\001 \002(\r\022\025\n\rlayou" + 22774 "tVersion\030\002 \002(\r\022\r\n\005codec\030\003 \001(\t\022:\n\010section" + 22775 "s\030\004 \003(\0132(.hadoop.hdfs.fsimage.FileSummar" + 22776 "y.Section\0327\n\007Section\022\014\n\004name\030\001 \001(\t\022\016\n\006le" + 22777 "ngth\030\002 \001(\004\022\016\n\006offset\030\003 \001(\004\"\277\001\n\021NameSyste" + 22778 "mSection\022\023\n\013namespaceId\030\001 \001(\r\022\022\n\ngenstam" + 22779 "pV1\030\002 \001(\004\022\022\n\ngenstampV2\030\003 \001(\004\022\027\n\017genstam" + 22780 "pV1Limit\030\004 \001(\004\022\034\n\024lastAllocatedBlockId\030\005", 22781 " \001(\004\022\025\n\rtransactionId\030\006 \001(\004\022\037\n\027rollingUp" + 22782 "gradeStartTime\030\007 \001(\004\"\346\n\n\014INodeSection\022\023\n" + 22783 "\013lastInodeId\030\001 \001(\004\022\021\n\tnumInodes\030\002 \001(\004\032I\n" + 22784 "\034FileUnderConstructionFeature\022\022\n\nclientN" + 22785 "ame\030\001 \001(\t\022\025\n\rclientMachine\030\002 \001(\t\032&\n\017AclF" + 22786 "eatureProto\022\023\n\007entries\030\002 \003(\007B\002\020\001\0320\n\021XAtt" + 22787 "rCompactProto\022\014\n\004name\030\001 \002(\007\022\r\n\005value\030\002 \001" + 22788 "(\014\032X\n\021XAttrFeatureProto\022C\n\006xAttrs\030\001 \003(\0132" + 22789 "3.hadoop.hdfs.fsimage.INodeSection.XAttr" + 22790 "CompactProto\032\225\003\n\tINodeFile\022\023\n\013replicatio", 22791 "n\030\001 \001(\r\022\030\n\020modificationTime\030\002 \001(\004\022\022\n\nacc" + 22792 "essTime\030\003 \001(\004\022\032\n\022preferredBlockSize\030\004 \001(" + 22793 "\004\022\022\n\npermission\030\005 \001(\006\022\'\n\006blocks\030\006 \003(\0132\027." + 22794 "hadoop.hdfs.BlockProto\022N\n\006fileUC\030\007 \001(\0132>" + 22795 ".hadoop.hdfs.fsimage.INodeSection.FileUn" + 22796 "derConstructionFeature\022>\n\003acl\030\010 \001(\01321.ha" + 22797 "doop.hdfs.fsimage.INodeSection.AclFeatur" + 22798 "eProto\022C\n\006xAttrs\030\t \001(\01323.hadoop.hdfs.fsi" + 22799 "mage.INodeSection.XAttrFeatureProto\022\027\n\017s" + 22800 "toragePolicyID\030\n \001(\r\032\345\001\n\016INodeDirectory\022", 22801 "\030\n\020modificationTime\030\001 \001(\004\022\017\n\007nsQuota\030\002 \001" + 22802 "(\004\022\017\n\007dsQuota\030\003 \001(\004\022\022\n\npermission\030\004 \001(\006\022" + 22803 ">\n\003acl\030\005 \001(\01321.hadoop.hdfs.fsimage.INode" + 22804 "Section.AclFeatureProto\022C\n\006xAttrs\030\006 \001(\0132" + 22805 "3.hadoop.hdfs.fsimage.INodeSection.XAttr" + 22806 "FeatureProto\032`\n\014INodeSymlink\022\022\n\npermissi" + 22807 "on\030\001 \001(\006\022\016\n\006target\030\002 \001(\014\022\030\n\020modification" + 22808 "Time\030\003 \001(\004\022\022\n\naccessTime\030\004 \001(\004\032\314\002\n\005INode" + 22809 "\022:\n\004type\030\001 \002(\0162,.hadoop.hdfs.fsimage.INo" + 22810 "deSection.INode.Type\022\n\n\002id\030\002 \002(\004\022\014\n\004name", 22811 "\030\003 \001(\014\0229\n\004file\030\004 \001(\0132+.hadoop.hdfs.fsima" + 22812 "ge.INodeSection.INodeFile\022C\n\tdirectory\030\005" + 22813 " \001(\01320.hadoop.hdfs.fsimage.INodeSection." + 22814 "INodeDirectory\022?\n\007symlink\030\006 \001(\0132..hadoop" + 22815 ".hdfs.fsimage.INodeSection.INodeSymlink\"" + 22816 ",\n\004Type\022\010\n\004FILE\020\001\022\r\n\tDIRECTORY\020\002\022\013\n\007SYML" + 22817 "INK\020\003\"`\n\035FilesUnderConstructionSection\032?" + 22818 "\n\032FileUnderConstructionEntry\022\017\n\007inodeId\030" + 22819 "\001 \001(\004\022\020\n\010fullPath\030\002 \001(\t\"b\n\025INodeDirector" + 22820 "ySection\032I\n\010DirEntry\022\016\n\006parent\030\001 \001(\004\022\024\n\010", 22821 "children\030\002 \003(\004B\002\020\001\022\027\n\013refChildren\030\003 \003(\rB" + 22822 "\002\020\001\"z\n\025INodeReferenceSection\032a\n\016INodeRef" + 22823 "erence\022\022\n\nreferredId\030\001 \001(\004\022\014\n\004name\030\002 \001(\014" + 22824 "\022\025\n\rdstSnapshotId\030\003 \001(\r\022\026\n\016lastSnapshotI" + 22825 "d\030\004 \001(\r\"\265\001\n\017SnapshotSection\022\027\n\017snapshotC" + 22826 "ounter\030\001 \001(\r\022\034\n\020snapshottableDir\030\002 \003(\004B\002" + 22827 "\020\001\022\024\n\014numSnapshots\030\003 \001(\r\032U\n\010Snapshot\022\022\n\n" + 22828 "snapshotId\030\001 \001(\r\0225\n\004root\030\002 \001(\0132\'.hadoop." + 22829 "hdfs.fsimage.INodeSection.INode\"\327\004\n\023Snap" + 22830 "shotDiffSection\032 \n\020CreatedListEntry\022\014\n\004n", 22831 "ame\030\001 \001(\014\032\367\001\n\rDirectoryDiff\022\022\n\nsnapshotI" + 22832 "d\030\001 \001(\r\022\024\n\014childrenSize\030\002 \001(\r\022\026\n\016isSnaps" + 22833 "hotRoot\030\003 \001(\010\022\014\n\004name\030\004 \001(\014\022F\n\014snapshotC" + 22834 "opy\030\005 \001(\01320.hadoop.hdfs.fsimage.INodeSec" + 22835 "tion.INodeDirectory\022\027\n\017createdListSize\030\006" + 22836 " \001(\r\022\030\n\014deletedINode\030\007 \003(\004B\002\020\001\022\033\n\017delete" + 22837 "dINodeRef\030\010 \003(\rB\002\020\001\032\201\001\n\010FileDiff\022\022\n\nsnap" + 22838 "shotId\030\001 \001(\r\022\020\n\010fileSize\030\002 \001(\004\022\014\n\004name\030\003" + 22839 " \001(\014\022A\n\014snapshotCopy\030\004 \001(\0132+.hadoop.hdfs" + 22840 ".fsimage.INodeSection.INodeFile\032\237\001\n\tDiff", 22841 "Entry\022E\n\004type\030\001 \002(\01627.hadoop.hdfs.fsimag" + 22842 "e.SnapshotDiffSection.DiffEntry.Type\022\017\n\007" + 22843 "inodeId\030\002 \001(\004\022\021\n\tnumOfDiff\030\003 \001(\r\"\'\n\004Type" + 22844 "\022\014\n\010FILEDIFF\020\001\022\021\n\rDIRECTORYDIFF\020\002\"H\n\022Str" + 22845 "ingTableSection\022\020\n\010numEntry\030\001 \001(\r\032 \n\005Ent" + 22846 "ry\022\n\n\002id\030\001 \001(\r\022\013\n\003str\030\002 \001(\t\"\341\002\n\024SecretMa" + 22847 "nagerSection\022\021\n\tcurrentId\030\001 \001(\r\022\033\n\023token" + 22848 "SequenceNumber\030\002 \001(\r\022\017\n\007numKeys\030\003 \001(\r\022\021\n" + 22849 "\tnumTokens\030\004 \001(\r\032<\n\rDelegationKey\022\n\n\002id\030" + 22850 "\001 \001(\r\022\022\n\nexpiryDate\030\002 \001(\004\022\013\n\003key\030\003 \001(\014\032\266", 22851 "\001\n\014PersistToken\022\017\n\007version\030\001 \001(\r\022\r\n\005owne" + 22852 "r\030\002 \001(\t\022\017\n\007renewer\030\003 \001(\t\022\020\n\010realUser\030\004 \001" + 22853 "(\t\022\021\n\tissueDate\030\005 \001(\004\022\017\n\007maxDate\030\006 \001(\004\022\026" + 22854 "\n\016sequenceNumber\030\007 \001(\r\022\023\n\013masterKeyId\030\010 " + 22855 "\001(\r\022\022\n\nexpiryDate\030\t \001(\004\"W\n\023CacheManagerS" + 22856 "ection\022\027\n\017nextDirectiveId\030\001 \002(\004\022\020\n\010numPo" + 22857 "ols\030\002 \002(\r\022\025\n\rnumDirectives\030\003 \002(\rB6\n&org." + 22858 "apache.hadoop.hdfs.server.namenodeB\014FsIm" + 22859 "ageProto" 22860 }; 22861 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 22862 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 22863 public com.google.protobuf.ExtensionRegistry assignDescriptors( 22864 com.google.protobuf.Descriptors.FileDescriptor root) { 22865 descriptor = root; 22866 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor = 22867 getDescriptor().getMessageTypes().get(0); 22868 internal_static_hadoop_hdfs_fsimage_FileSummary_fieldAccessorTable = new 22869 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22870 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor, 22871 new java.lang.String[] { "OndiskVersion", "LayoutVersion", "Codec", "Sections", }); 22872 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor = 22873 internal_static_hadoop_hdfs_fsimage_FileSummary_descriptor.getNestedTypes().get(0); 22874 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_fieldAccessorTable = new 22875 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22876 internal_static_hadoop_hdfs_fsimage_FileSummary_Section_descriptor, 22877 new java.lang.String[] { "Name", "Length", "Offset", }); 22878 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor = 22879 getDescriptor().getMessageTypes().get(1); 22880 internal_static_hadoop_hdfs_fsimage_NameSystemSection_fieldAccessorTable = new 22881 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22882 internal_static_hadoop_hdfs_fsimage_NameSystemSection_descriptor, 22883 new java.lang.String[] { "NamespaceId", "GenstampV1", "GenstampV2", "GenstampV1Limit", "LastAllocatedBlockId", "TransactionId", "RollingUpgradeStartTime", }); 22884 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor = 22885 getDescriptor().getMessageTypes().get(2); 22886 internal_static_hadoop_hdfs_fsimage_INodeSection_fieldAccessorTable = new 22887 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22888 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor, 22889 new java.lang.String[] { "LastInodeId", "NumInodes", }); 22890 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor = 22891 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(0); 22892 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_fieldAccessorTable = new 22893 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22894 internal_static_hadoop_hdfs_fsimage_INodeSection_FileUnderConstructionFeature_descriptor, 22895 new java.lang.String[] { "ClientName", "ClientMachine", }); 22896 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor = 22897 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(1); 22898 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_fieldAccessorTable = new 22899 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22900 internal_static_hadoop_hdfs_fsimage_INodeSection_AclFeatureProto_descriptor, 22901 new java.lang.String[] { "Entries", }); 22902 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor = 22903 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(2); 22904 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_fieldAccessorTable = new 22905 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22906 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrCompactProto_descriptor, 22907 new java.lang.String[] { "Name", "Value", }); 22908 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor = 22909 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(3); 22910 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_fieldAccessorTable = new 22911 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22912 internal_static_hadoop_hdfs_fsimage_INodeSection_XAttrFeatureProto_descriptor, 22913 new java.lang.String[] { "XAttrs", }); 22914 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor = 22915 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(4); 22916 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_fieldAccessorTable = new 22917 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22918 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeFile_descriptor, 22919 new java.lang.String[] { "Replication", "ModificationTime", "AccessTime", "PreferredBlockSize", "Permission", "Blocks", "FileUC", "Acl", "XAttrs", "StoragePolicyID", }); 22920 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor = 22921 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(5); 22922 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_fieldAccessorTable = new 22923 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22924 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeDirectory_descriptor, 22925 new java.lang.String[] { "ModificationTime", "NsQuota", "DsQuota", "Permission", "Acl", "XAttrs", }); 22926 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor = 22927 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(6); 22928 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_fieldAccessorTable = new 22929 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22930 internal_static_hadoop_hdfs_fsimage_INodeSection_INodeSymlink_descriptor, 22931 new java.lang.String[] { "Permission", "Target", "ModificationTime", "AccessTime", }); 22932 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor = 22933 internal_static_hadoop_hdfs_fsimage_INodeSection_descriptor.getNestedTypes().get(7); 22934 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_fieldAccessorTable = new 22935 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22936 internal_static_hadoop_hdfs_fsimage_INodeSection_INode_descriptor, 22937 new java.lang.String[] { "Type", "Id", "Name", "File", "Directory", "Symlink", }); 22938 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor = 22939 getDescriptor().getMessageTypes().get(3); 22940 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_fieldAccessorTable = new 22941 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22942 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor, 22943 new java.lang.String[] { }); 22944 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor = 22945 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_descriptor.getNestedTypes().get(0); 22946 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_fieldAccessorTable = new 22947 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22948 internal_static_hadoop_hdfs_fsimage_FilesUnderConstructionSection_FileUnderConstructionEntry_descriptor, 22949 new java.lang.String[] { "InodeId", "FullPath", }); 22950 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor = 22951 getDescriptor().getMessageTypes().get(4); 22952 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_fieldAccessorTable = new 22953 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22954 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor, 22955 new java.lang.String[] { }); 22956 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor = 22957 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_descriptor.getNestedTypes().get(0); 22958 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_fieldAccessorTable = new 22959 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22960 internal_static_hadoop_hdfs_fsimage_INodeDirectorySection_DirEntry_descriptor, 22961 new java.lang.String[] { "Parent", "Children", "RefChildren", }); 22962 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor = 22963 getDescriptor().getMessageTypes().get(5); 22964 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_fieldAccessorTable = new 22965 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22966 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor, 22967 new java.lang.String[] { }); 22968 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor = 22969 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_descriptor.getNestedTypes().get(0); 22970 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_fieldAccessorTable = new 22971 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22972 internal_static_hadoop_hdfs_fsimage_INodeReferenceSection_INodeReference_descriptor, 22973 new java.lang.String[] { "ReferredId", "Name", "DstSnapshotId", "LastSnapshotId", }); 22974 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor = 22975 getDescriptor().getMessageTypes().get(6); 22976 internal_static_hadoop_hdfs_fsimage_SnapshotSection_fieldAccessorTable = new 22977 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22978 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor, 22979 new java.lang.String[] { "SnapshotCounter", "SnapshottableDir", "NumSnapshots", }); 22980 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor = 22981 internal_static_hadoop_hdfs_fsimage_SnapshotSection_descriptor.getNestedTypes().get(0); 22982 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_fieldAccessorTable = new 22983 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22984 internal_static_hadoop_hdfs_fsimage_SnapshotSection_Snapshot_descriptor, 22985 new java.lang.String[] { "SnapshotId", "Root", }); 22986 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor = 22987 getDescriptor().getMessageTypes().get(7); 22988 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_fieldAccessorTable = new 22989 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22990 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor, 22991 new java.lang.String[] { }); 22992 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor = 22993 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(0); 22994 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_fieldAccessorTable = new 22995 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 22996 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_CreatedListEntry_descriptor, 22997 new java.lang.String[] { "Name", }); 22998 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor = 22999 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(1); 23000 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_fieldAccessorTable = new 23001 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23002 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DirectoryDiff_descriptor, 23003 new java.lang.String[] { "SnapshotId", "ChildrenSize", "IsSnapshotRoot", "Name", "SnapshotCopy", "CreatedListSize", "DeletedINode", "DeletedINodeRef", }); 23004 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor = 23005 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(2); 23006 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_fieldAccessorTable = new 23007 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23008 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_FileDiff_descriptor, 23009 new java.lang.String[] { "SnapshotId", "FileSize", "Name", "SnapshotCopy", }); 23010 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor = 23011 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_descriptor.getNestedTypes().get(3); 23012 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_fieldAccessorTable = new 23013 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23014 internal_static_hadoop_hdfs_fsimage_SnapshotDiffSection_DiffEntry_descriptor, 23015 new java.lang.String[] { "Type", "InodeId", "NumOfDiff", }); 23016 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor = 23017 getDescriptor().getMessageTypes().get(8); 23018 internal_static_hadoop_hdfs_fsimage_StringTableSection_fieldAccessorTable = new 23019 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23020 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor, 23021 new java.lang.String[] { "NumEntry", }); 23022 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor = 23023 internal_static_hadoop_hdfs_fsimage_StringTableSection_descriptor.getNestedTypes().get(0); 23024 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_fieldAccessorTable = new 23025 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23026 internal_static_hadoop_hdfs_fsimage_StringTableSection_Entry_descriptor, 23027 new java.lang.String[] { "Id", "Str", }); 23028 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor = 23029 getDescriptor().getMessageTypes().get(9); 23030 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_fieldAccessorTable = new 23031 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23032 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor, 23033 new java.lang.String[] { "CurrentId", "TokenSequenceNumber", "NumKeys", "NumTokens", }); 23034 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor = 23035 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor.getNestedTypes().get(0); 23036 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_fieldAccessorTable = new 23037 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23038 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_DelegationKey_descriptor, 23039 new java.lang.String[] { "Id", "ExpiryDate", "Key", }); 23040 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor = 23041 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_descriptor.getNestedTypes().get(1); 23042 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_fieldAccessorTable = new 23043 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23044 internal_static_hadoop_hdfs_fsimage_SecretManagerSection_PersistToken_descriptor, 23045 new java.lang.String[] { "Version", "Owner", "Renewer", "RealUser", "IssueDate", "MaxDate", "SequenceNumber", "MasterKeyId", "ExpiryDate", }); 23046 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor = 23047 getDescriptor().getMessageTypes().get(10); 23048 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_fieldAccessorTable = new 23049 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 23050 internal_static_hadoop_hdfs_fsimage_CacheManagerSection_descriptor, 23051 new java.lang.String[] { "NextDirectiveId", "NumPools", "NumDirectives", }); 23052 return null; 23053 } 23054 }; 23055 com.google.protobuf.Descriptors.FileDescriptor 23056 .internalBuildGeneratedFileFrom(descriptorData, 23057 new com.google.protobuf.Descriptors.FileDescriptor[] { 23058 org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(), 23059 org.apache.hadoop.hdfs.protocol.proto.AclProtos.getDescriptor(), 23060 org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.getDescriptor(), 23061 }, assigner); 23062 } 23063 23064 // @@protoc_insertion_point(outer_class_scope) 23065 }