1 // Generated by the protocol buffer compiler. DO NOT EDIT! 2 // source: Master.proto 3 4 package org.apache.hadoop.hbase.protobuf.generated; 5 6 public final class MasterProtos { MasterProtos()7 private MasterProtos() {} registerAllExtensions( com.google.protobuf.ExtensionRegistry registry)8 public static void registerAllExtensions( 9 com.google.protobuf.ExtensionRegistry registry) { 10 } 11 public interface AddColumnRequestOrBuilder 12 extends com.google.protobuf.MessageOrBuilder { 13 14 // required .TableName table_name = 1; 15 /** 16 * <code>required .TableName table_name = 1;</code> 17 */ hasTableName()18 boolean hasTableName(); 19 /** 20 * <code>required .TableName table_name = 1;</code> 21 */ getTableName()22 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); 23 /** 24 * <code>required .TableName table_name = 1;</code> 25 */ getTableNameOrBuilder()26 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); 27 28 // required .ColumnFamilySchema column_families = 2; 29 /** 30 * <code>required .ColumnFamilySchema column_families = 2;</code> 31 */ hasColumnFamilies()32 boolean hasColumnFamilies(); 33 /** 34 * <code>required .ColumnFamilySchema column_families = 2;</code> 35 */ getColumnFamilies()36 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); 37 /** 38 * <code>required .ColumnFamilySchema column_families = 2;</code> 39 */ getColumnFamiliesOrBuilder()40 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); 41 42 // optional uint64 nonce_group = 3 [default = 0]; 43 /** 44 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 45 */ hasNonceGroup()46 boolean hasNonceGroup(); 47 /** 48 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 49 */ getNonceGroup()50 long getNonceGroup(); 51 52 // optional uint64 nonce = 4 [default = 0]; 53 /** 54 * <code>optional uint64 nonce = 4 [default = 0];</code> 55 */ hasNonce()56 boolean hasNonce(); 57 /** 58 * <code>optional uint64 nonce = 4 [default = 0];</code> 59 */ getNonce()60 long getNonce(); 61 } 62 /** 63 * Protobuf type {@code AddColumnRequest} 64 */ 65 public static final class AddColumnRequest extends 66 com.google.protobuf.GeneratedMessage 67 implements AddColumnRequestOrBuilder { 68 // Use AddColumnRequest.newBuilder() to construct. AddColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)69 private AddColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 70 super(builder); 71 this.unknownFields = builder.getUnknownFields(); 72 } AddColumnRequest(boolean noInit)73 private AddColumnRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 74 75 private static final AddColumnRequest defaultInstance; getDefaultInstance()76 public static AddColumnRequest getDefaultInstance() { 77 return defaultInstance; 78 } 79 getDefaultInstanceForType()80 public AddColumnRequest getDefaultInstanceForType() { 81 return defaultInstance; 82 } 83 84 private final com.google.protobuf.UnknownFieldSet unknownFields; 85 @java.lang.Override 86 public final com.google.protobuf.UnknownFieldSet getUnknownFields()87 getUnknownFields() { 88 return this.unknownFields; 89 } AddColumnRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)90 private AddColumnRequest( 91 com.google.protobuf.CodedInputStream input, 92 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 93 throws com.google.protobuf.InvalidProtocolBufferException { 94 initFields(); 95 int mutable_bitField0_ = 0; 96 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 97 com.google.protobuf.UnknownFieldSet.newBuilder(); 98 try { 99 boolean done = false; 100 while (!done) { 101 int tag = input.readTag(); 102 switch (tag) { 103 case 0: 104 done = true; 105 break; 106 default: { 107 if (!parseUnknownField(input, unknownFields, 108 extensionRegistry, tag)) { 109 done = true; 110 } 111 break; 112 } 113 case 10: { 114 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; 115 if (((bitField0_ & 0x00000001) == 0x00000001)) { 116 subBuilder = tableName_.toBuilder(); 117 } 118 tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); 119 if (subBuilder != null) { 120 subBuilder.mergeFrom(tableName_); 121 tableName_ = subBuilder.buildPartial(); 122 } 123 bitField0_ |= 0x00000001; 124 break; 125 } 126 case 18: { 127 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = null; 128 if (((bitField0_ & 0x00000002) == 0x00000002)) { 129 subBuilder = columnFamilies_.toBuilder(); 130 } 131 columnFamilies_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry); 132 if (subBuilder != null) { 133 subBuilder.mergeFrom(columnFamilies_); 134 columnFamilies_ = subBuilder.buildPartial(); 135 } 136 bitField0_ |= 0x00000002; 137 break; 138 } 139 case 24: { 140 bitField0_ |= 0x00000004; 141 nonceGroup_ = input.readUInt64(); 142 break; 143 } 144 case 32: { 145 bitField0_ |= 0x00000008; 146 nonce_ = input.readUInt64(); 147 break; 148 } 149 } 150 } 151 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 152 throw e.setUnfinishedMessage(this); 153 } catch (java.io.IOException e) { 154 throw new com.google.protobuf.InvalidProtocolBufferException( 155 e.getMessage()).setUnfinishedMessage(this); 156 } finally { 157 this.unknownFields = unknownFields.build(); 158 makeExtensionsImmutable(); 159 } 160 } 161 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()162 getDescriptor() { 163 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnRequest_descriptor; 164 } 165 166 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()167 internalGetFieldAccessorTable() { 168 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnRequest_fieldAccessorTable 169 .ensureFieldAccessorsInitialized( 170 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.Builder.class); 171 } 172 173 public static com.google.protobuf.Parser<AddColumnRequest> PARSER = 174 new com.google.protobuf.AbstractParser<AddColumnRequest>() { 175 public AddColumnRequest parsePartialFrom( 176 com.google.protobuf.CodedInputStream input, 177 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 178 throws com.google.protobuf.InvalidProtocolBufferException { 179 return new AddColumnRequest(input, extensionRegistry); 180 } 181 }; 182 183 @java.lang.Override getParserForType()184 public com.google.protobuf.Parser<AddColumnRequest> getParserForType() { 185 return PARSER; 186 } 187 188 private int bitField0_; 189 // required .TableName table_name = 1; 190 public static final int TABLE_NAME_FIELD_NUMBER = 1; 191 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; 192 /** 193 * <code>required .TableName table_name = 1;</code> 194 */ hasTableName()195 public boolean hasTableName() { 196 return ((bitField0_ & 0x00000001) == 0x00000001); 197 } 198 /** 199 * <code>required .TableName table_name = 1;</code> 200 */ getTableName()201 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 202 return tableName_; 203 } 204 /** 205 * <code>required .TableName table_name = 1;</code> 206 */ getTableNameOrBuilder()207 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 208 return tableName_; 209 } 210 211 // required .ColumnFamilySchema column_families = 2; 212 public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2; 213 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; 214 /** 215 * <code>required .ColumnFamilySchema column_families = 2;</code> 216 */ hasColumnFamilies()217 public boolean hasColumnFamilies() { 218 return ((bitField0_ & 0x00000002) == 0x00000002); 219 } 220 /** 221 * <code>required .ColumnFamilySchema column_families = 2;</code> 222 */ getColumnFamilies()223 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() { 224 return columnFamilies_; 225 } 226 /** 227 * <code>required .ColumnFamilySchema column_families = 2;</code> 228 */ getColumnFamiliesOrBuilder()229 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() { 230 return columnFamilies_; 231 } 232 233 // optional uint64 nonce_group = 3 [default = 0]; 234 public static final int NONCE_GROUP_FIELD_NUMBER = 3; 235 private long nonceGroup_; 236 /** 237 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 238 */ hasNonceGroup()239 public boolean hasNonceGroup() { 240 return ((bitField0_ & 0x00000004) == 0x00000004); 241 } 242 /** 243 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 244 */ getNonceGroup()245 public long getNonceGroup() { 246 return nonceGroup_; 247 } 248 249 // optional uint64 nonce = 4 [default = 0]; 250 public static final int NONCE_FIELD_NUMBER = 4; 251 private long nonce_; 252 /** 253 * <code>optional uint64 nonce = 4 [default = 0];</code> 254 */ hasNonce()255 public boolean hasNonce() { 256 return ((bitField0_ & 0x00000008) == 0x00000008); 257 } 258 /** 259 * <code>optional uint64 nonce = 4 [default = 0];</code> 260 */ getNonce()261 public long getNonce() { 262 return nonce_; 263 } 264 initFields()265 private void initFields() { 266 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 267 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 268 nonceGroup_ = 0L; 269 nonce_ = 0L; 270 } 271 private byte memoizedIsInitialized = -1; isInitialized()272 public final boolean isInitialized() { 273 byte isInitialized = memoizedIsInitialized; 274 if (isInitialized != -1) return isInitialized == 1; 275 276 if (!hasTableName()) { 277 memoizedIsInitialized = 0; 278 return false; 279 } 280 if (!hasColumnFamilies()) { 281 memoizedIsInitialized = 0; 282 return false; 283 } 284 if (!getTableName().isInitialized()) { 285 memoizedIsInitialized = 0; 286 return false; 287 } 288 if (!getColumnFamilies().isInitialized()) { 289 memoizedIsInitialized = 0; 290 return false; 291 } 292 memoizedIsInitialized = 1; 293 return true; 294 } 295 writeTo(com.google.protobuf.CodedOutputStream output)296 public void writeTo(com.google.protobuf.CodedOutputStream output) 297 throws java.io.IOException { 298 getSerializedSize(); 299 if (((bitField0_ & 0x00000001) == 0x00000001)) { 300 output.writeMessage(1, tableName_); 301 } 302 if (((bitField0_ & 0x00000002) == 0x00000002)) { 303 output.writeMessage(2, columnFamilies_); 304 } 305 if (((bitField0_ & 0x00000004) == 0x00000004)) { 306 output.writeUInt64(3, nonceGroup_); 307 } 308 if (((bitField0_ & 0x00000008) == 0x00000008)) { 309 output.writeUInt64(4, nonce_); 310 } 311 getUnknownFields().writeTo(output); 312 } 313 314 private int memoizedSerializedSize = -1; getSerializedSize()315 public int getSerializedSize() { 316 int size = memoizedSerializedSize; 317 if (size != -1) return size; 318 319 size = 0; 320 if (((bitField0_ & 0x00000001) == 0x00000001)) { 321 size += com.google.protobuf.CodedOutputStream 322 .computeMessageSize(1, tableName_); 323 } 324 if (((bitField0_ & 0x00000002) == 0x00000002)) { 325 size += com.google.protobuf.CodedOutputStream 326 .computeMessageSize(2, columnFamilies_); 327 } 328 if (((bitField0_ & 0x00000004) == 0x00000004)) { 329 size += com.google.protobuf.CodedOutputStream 330 .computeUInt64Size(3, nonceGroup_); 331 } 332 if (((bitField0_ & 0x00000008) == 0x00000008)) { 333 size += com.google.protobuf.CodedOutputStream 334 .computeUInt64Size(4, nonce_); 335 } 336 size += getUnknownFields().getSerializedSize(); 337 memoizedSerializedSize = size; 338 return size; 339 } 340 341 private static final long serialVersionUID = 0L; 342 @java.lang.Override writeReplace()343 protected java.lang.Object writeReplace() 344 throws java.io.ObjectStreamException { 345 return super.writeReplace(); 346 } 347 348 @java.lang.Override equals(final java.lang.Object obj)349 public boolean equals(final java.lang.Object obj) { 350 if (obj == this) { 351 return true; 352 } 353 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest)) { 354 return super.equals(obj); 355 } 356 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest) obj; 357 358 boolean result = true; 359 result = result && (hasTableName() == other.hasTableName()); 360 if (hasTableName()) { 361 result = result && getTableName() 362 .equals(other.getTableName()); 363 } 364 result = result && (hasColumnFamilies() == other.hasColumnFamilies()); 365 if (hasColumnFamilies()) { 366 result = result && getColumnFamilies() 367 .equals(other.getColumnFamilies()); 368 } 369 result = result && (hasNonceGroup() == other.hasNonceGroup()); 370 if (hasNonceGroup()) { 371 result = result && (getNonceGroup() 372 == other.getNonceGroup()); 373 } 374 result = result && (hasNonce() == other.hasNonce()); 375 if (hasNonce()) { 376 result = result && (getNonce() 377 == other.getNonce()); 378 } 379 result = result && 380 getUnknownFields().equals(other.getUnknownFields()); 381 return result; 382 } 383 384 private int memoizedHashCode = 0; 385 @java.lang.Override hashCode()386 public int hashCode() { 387 if (memoizedHashCode != 0) { 388 return memoizedHashCode; 389 } 390 int hash = 41; 391 hash = (19 * hash) + getDescriptorForType().hashCode(); 392 if (hasTableName()) { 393 hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; 394 hash = (53 * hash) + getTableName().hashCode(); 395 } 396 if (hasColumnFamilies()) { 397 hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER; 398 hash = (53 * hash) + getColumnFamilies().hashCode(); 399 } 400 if (hasNonceGroup()) { 401 hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; 402 hash = (53 * hash) + hashLong(getNonceGroup()); 403 } 404 if (hasNonce()) { 405 hash = (37 * hash) + NONCE_FIELD_NUMBER; 406 hash = (53 * hash) + hashLong(getNonce()); 407 } 408 hash = (29 * hash) + getUnknownFields().hashCode(); 409 memoizedHashCode = hash; 410 return hash; 411 } 412 parseFrom( com.google.protobuf.ByteString data)413 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 414 com.google.protobuf.ByteString data) 415 throws com.google.protobuf.InvalidProtocolBufferException { 416 return PARSER.parseFrom(data); 417 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)418 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 419 com.google.protobuf.ByteString data, 420 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 421 throws com.google.protobuf.InvalidProtocolBufferException { 422 return PARSER.parseFrom(data, extensionRegistry); 423 } parseFrom(byte[] data)424 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(byte[] data) 425 throws com.google.protobuf.InvalidProtocolBufferException { 426 return PARSER.parseFrom(data); 427 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)428 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 429 byte[] data, 430 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 431 throws com.google.protobuf.InvalidProtocolBufferException { 432 return PARSER.parseFrom(data, extensionRegistry); 433 } parseFrom(java.io.InputStream input)434 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(java.io.InputStream input) 435 throws java.io.IOException { 436 return PARSER.parseFrom(input); 437 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)438 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 439 java.io.InputStream input, 440 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 441 throws java.io.IOException { 442 return PARSER.parseFrom(input, extensionRegistry); 443 } parseDelimitedFrom(java.io.InputStream input)444 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom(java.io.InputStream input) 445 throws java.io.IOException { 446 return PARSER.parseDelimitedFrom(input); 447 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)448 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom( 449 java.io.InputStream input, 450 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 451 throws java.io.IOException { 452 return PARSER.parseDelimitedFrom(input, extensionRegistry); 453 } parseFrom( com.google.protobuf.CodedInputStream input)454 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 455 com.google.protobuf.CodedInputStream input) 456 throws java.io.IOException { 457 return PARSER.parseFrom(input); 458 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)459 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parseFrom( 460 com.google.protobuf.CodedInputStream input, 461 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 462 throws java.io.IOException { 463 return PARSER.parseFrom(input, extensionRegistry); 464 } 465 newBuilder()466 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()467 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest prototype)468 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest prototype) { 469 return newBuilder().mergeFrom(prototype); 470 } toBuilder()471 public Builder toBuilder() { return newBuilder(this); } 472 473 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)474 protected Builder newBuilderForType( 475 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 476 Builder builder = new Builder(parent); 477 return builder; 478 } 479 /** 480 * Protobuf type {@code AddColumnRequest} 481 */ 482 public static final class Builder extends 483 com.google.protobuf.GeneratedMessage.Builder<Builder> 484 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequestOrBuilder { 485 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()486 getDescriptor() { 487 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnRequest_descriptor; 488 } 489 490 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()491 internalGetFieldAccessorTable() { 492 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnRequest_fieldAccessorTable 493 .ensureFieldAccessorsInitialized( 494 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.Builder.class); 495 } 496 497 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.newBuilder() Builder()498 private Builder() { 499 maybeForceBuilderInitialization(); 500 } 501 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)502 private Builder( 503 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 504 super(parent); 505 maybeForceBuilderInitialization(); 506 } maybeForceBuilderInitialization()507 private void maybeForceBuilderInitialization() { 508 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 509 getTableNameFieldBuilder(); 510 getColumnFamiliesFieldBuilder(); 511 } 512 } create()513 private static Builder create() { 514 return new Builder(); 515 } 516 clear()517 public Builder clear() { 518 super.clear(); 519 if (tableNameBuilder_ == null) { 520 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 521 } else { 522 tableNameBuilder_.clear(); 523 } 524 bitField0_ = (bitField0_ & ~0x00000001); 525 if (columnFamiliesBuilder_ == null) { 526 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 527 } else { 528 columnFamiliesBuilder_.clear(); 529 } 530 bitField0_ = (bitField0_ & ~0x00000002); 531 nonceGroup_ = 0L; 532 bitField0_ = (bitField0_ & ~0x00000004); 533 nonce_ = 0L; 534 bitField0_ = (bitField0_ & ~0x00000008); 535 return this; 536 } 537 clone()538 public Builder clone() { 539 return create().mergeFrom(buildPartial()); 540 } 541 542 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()543 getDescriptorForType() { 544 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnRequest_descriptor; 545 } 546 getDefaultInstanceForType()547 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest getDefaultInstanceForType() { 548 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance(); 549 } 550 build()551 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest build() { 552 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest result = buildPartial(); 553 if (!result.isInitialized()) { 554 throw newUninitializedMessageException(result); 555 } 556 return result; 557 } 558 buildPartial()559 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest buildPartial() { 560 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest(this); 561 int from_bitField0_ = bitField0_; 562 int to_bitField0_ = 0; 563 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 564 to_bitField0_ |= 0x00000001; 565 } 566 if (tableNameBuilder_ == null) { 567 result.tableName_ = tableName_; 568 } else { 569 result.tableName_ = tableNameBuilder_.build(); 570 } 571 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 572 to_bitField0_ |= 0x00000002; 573 } 574 if (columnFamiliesBuilder_ == null) { 575 result.columnFamilies_ = columnFamilies_; 576 } else { 577 result.columnFamilies_ = columnFamiliesBuilder_.build(); 578 } 579 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 580 to_bitField0_ |= 0x00000004; 581 } 582 result.nonceGroup_ = nonceGroup_; 583 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 584 to_bitField0_ |= 0x00000008; 585 } 586 result.nonce_ = nonce_; 587 result.bitField0_ = to_bitField0_; 588 onBuilt(); 589 return result; 590 } 591 mergeFrom(com.google.protobuf.Message other)592 public Builder mergeFrom(com.google.protobuf.Message other) { 593 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest) { 594 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest)other); 595 } else { 596 super.mergeFrom(other); 597 return this; 598 } 599 } 600 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest other)601 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest other) { 602 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance()) return this; 603 if (other.hasTableName()) { 604 mergeTableName(other.getTableName()); 605 } 606 if (other.hasColumnFamilies()) { 607 mergeColumnFamilies(other.getColumnFamilies()); 608 } 609 if (other.hasNonceGroup()) { 610 setNonceGroup(other.getNonceGroup()); 611 } 612 if (other.hasNonce()) { 613 setNonce(other.getNonce()); 614 } 615 this.mergeUnknownFields(other.getUnknownFields()); 616 return this; 617 } 618 isInitialized()619 public final boolean isInitialized() { 620 if (!hasTableName()) { 621 622 return false; 623 } 624 if (!hasColumnFamilies()) { 625 626 return false; 627 } 628 if (!getTableName().isInitialized()) { 629 630 return false; 631 } 632 if (!getColumnFamilies().isInitialized()) { 633 634 return false; 635 } 636 return true; 637 } 638 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)639 public Builder mergeFrom( 640 com.google.protobuf.CodedInputStream input, 641 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 642 throws java.io.IOException { 643 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest parsedMessage = null; 644 try { 645 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 646 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 647 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnRequest) e.getUnfinishedMessage(); 648 throw e; 649 } finally { 650 if (parsedMessage != null) { 651 mergeFrom(parsedMessage); 652 } 653 } 654 return this; 655 } 656 private int bitField0_; 657 658 // required .TableName table_name = 1; 659 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 660 private com.google.protobuf.SingleFieldBuilder< 661 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; 662 /** 663 * <code>required .TableName table_name = 1;</code> 664 */ hasTableName()665 public boolean hasTableName() { 666 return ((bitField0_ & 0x00000001) == 0x00000001); 667 } 668 /** 669 * <code>required .TableName table_name = 1;</code> 670 */ getTableName()671 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 672 if (tableNameBuilder_ == null) { 673 return tableName_; 674 } else { 675 return tableNameBuilder_.getMessage(); 676 } 677 } 678 /** 679 * <code>required .TableName table_name = 1;</code> 680 */ setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)681 public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 682 if (tableNameBuilder_ == null) { 683 if (value == null) { 684 throw new NullPointerException(); 685 } 686 tableName_ = value; 687 onChanged(); 688 } else { 689 tableNameBuilder_.setMessage(value); 690 } 691 bitField0_ |= 0x00000001; 692 return this; 693 } 694 /** 695 * <code>required .TableName table_name = 1;</code> 696 */ setTableName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue)697 public Builder setTableName( 698 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { 699 if (tableNameBuilder_ == null) { 700 tableName_ = builderForValue.build(); 701 onChanged(); 702 } else { 703 tableNameBuilder_.setMessage(builderForValue.build()); 704 } 705 bitField0_ |= 0x00000001; 706 return this; 707 } 708 /** 709 * <code>required .TableName table_name = 1;</code> 710 */ mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)711 public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 712 if (tableNameBuilder_ == null) { 713 if (((bitField0_ & 0x00000001) == 0x00000001) && 714 tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { 715 tableName_ = 716 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); 717 } else { 718 tableName_ = value; 719 } 720 onChanged(); 721 } else { 722 tableNameBuilder_.mergeFrom(value); 723 } 724 bitField0_ |= 0x00000001; 725 return this; 726 } 727 /** 728 * <code>required .TableName table_name = 1;</code> 729 */ clearTableName()730 public Builder clearTableName() { 731 if (tableNameBuilder_ == null) { 732 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 733 onChanged(); 734 } else { 735 tableNameBuilder_.clear(); 736 } 737 bitField0_ = (bitField0_ & ~0x00000001); 738 return this; 739 } 740 /** 741 * <code>required .TableName table_name = 1;</code> 742 */ getTableNameBuilder()743 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { 744 bitField0_ |= 0x00000001; 745 onChanged(); 746 return getTableNameFieldBuilder().getBuilder(); 747 } 748 /** 749 * <code>required .TableName table_name = 1;</code> 750 */ getTableNameOrBuilder()751 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 752 if (tableNameBuilder_ != null) { 753 return tableNameBuilder_.getMessageOrBuilder(); 754 } else { 755 return tableName_; 756 } 757 } 758 /** 759 * <code>required .TableName table_name = 1;</code> 760 */ 761 private com.google.protobuf.SingleFieldBuilder< 762 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder()763 getTableNameFieldBuilder() { 764 if (tableNameBuilder_ == null) { 765 tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< 766 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( 767 tableName_, 768 getParentForChildren(), 769 isClean()); 770 tableName_ = null; 771 } 772 return tableNameBuilder_; 773 } 774 775 // required .ColumnFamilySchema column_families = 2; 776 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 777 private com.google.protobuf.SingleFieldBuilder< 778 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; 779 /** 780 * <code>required .ColumnFamilySchema column_families = 2;</code> 781 */ hasColumnFamilies()782 public boolean hasColumnFamilies() { 783 return ((bitField0_ & 0x00000002) == 0x00000002); 784 } 785 /** 786 * <code>required .ColumnFamilySchema column_families = 2;</code> 787 */ getColumnFamilies()788 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() { 789 if (columnFamiliesBuilder_ == null) { 790 return columnFamilies_; 791 } else { 792 return columnFamiliesBuilder_.getMessage(); 793 } 794 } 795 /** 796 * <code>required .ColumnFamilySchema column_families = 2;</code> 797 */ setColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value)798 public Builder setColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { 799 if (columnFamiliesBuilder_ == null) { 800 if (value == null) { 801 throw new NullPointerException(); 802 } 803 columnFamilies_ = value; 804 onChanged(); 805 } else { 806 columnFamiliesBuilder_.setMessage(value); 807 } 808 bitField0_ |= 0x00000002; 809 return this; 810 } 811 /** 812 * <code>required .ColumnFamilySchema column_families = 2;</code> 813 */ setColumnFamilies( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue)814 public Builder setColumnFamilies( 815 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { 816 if (columnFamiliesBuilder_ == null) { 817 columnFamilies_ = builderForValue.build(); 818 onChanged(); 819 } else { 820 columnFamiliesBuilder_.setMessage(builderForValue.build()); 821 } 822 bitField0_ |= 0x00000002; 823 return this; 824 } 825 /** 826 * <code>required .ColumnFamilySchema column_families = 2;</code> 827 */ mergeColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value)828 public Builder mergeColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { 829 if (columnFamiliesBuilder_ == null) { 830 if (((bitField0_ & 0x00000002) == 0x00000002) && 831 columnFamilies_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) { 832 columnFamilies_ = 833 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(columnFamilies_).mergeFrom(value).buildPartial(); 834 } else { 835 columnFamilies_ = value; 836 } 837 onChanged(); 838 } else { 839 columnFamiliesBuilder_.mergeFrom(value); 840 } 841 bitField0_ |= 0x00000002; 842 return this; 843 } 844 /** 845 * <code>required .ColumnFamilySchema column_families = 2;</code> 846 */ clearColumnFamilies()847 public Builder clearColumnFamilies() { 848 if (columnFamiliesBuilder_ == null) { 849 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 850 onChanged(); 851 } else { 852 columnFamiliesBuilder_.clear(); 853 } 854 bitField0_ = (bitField0_ & ~0x00000002); 855 return this; 856 } 857 /** 858 * <code>required .ColumnFamilySchema column_families = 2;</code> 859 */ getColumnFamiliesBuilder()860 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder() { 861 bitField0_ |= 0x00000002; 862 onChanged(); 863 return getColumnFamiliesFieldBuilder().getBuilder(); 864 } 865 /** 866 * <code>required .ColumnFamilySchema column_families = 2;</code> 867 */ getColumnFamiliesOrBuilder()868 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() { 869 if (columnFamiliesBuilder_ != null) { 870 return columnFamiliesBuilder_.getMessageOrBuilder(); 871 } else { 872 return columnFamilies_; 873 } 874 } 875 /** 876 * <code>required .ColumnFamilySchema column_families = 2;</code> 877 */ 878 private com.google.protobuf.SingleFieldBuilder< 879 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> getColumnFamiliesFieldBuilder()880 getColumnFamiliesFieldBuilder() { 881 if (columnFamiliesBuilder_ == null) { 882 columnFamiliesBuilder_ = new com.google.protobuf.SingleFieldBuilder< 883 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>( 884 columnFamilies_, 885 getParentForChildren(), 886 isClean()); 887 columnFamilies_ = null; 888 } 889 return columnFamiliesBuilder_; 890 } 891 892 // optional uint64 nonce_group = 3 [default = 0]; 893 private long nonceGroup_ ; 894 /** 895 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 896 */ hasNonceGroup()897 public boolean hasNonceGroup() { 898 return ((bitField0_ & 0x00000004) == 0x00000004); 899 } 900 /** 901 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 902 */ getNonceGroup()903 public long getNonceGroup() { 904 return nonceGroup_; 905 } 906 /** 907 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 908 */ setNonceGroup(long value)909 public Builder setNonceGroup(long value) { 910 bitField0_ |= 0x00000004; 911 nonceGroup_ = value; 912 onChanged(); 913 return this; 914 } 915 /** 916 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 917 */ clearNonceGroup()918 public Builder clearNonceGroup() { 919 bitField0_ = (bitField0_ & ~0x00000004); 920 nonceGroup_ = 0L; 921 onChanged(); 922 return this; 923 } 924 925 // optional uint64 nonce = 4 [default = 0]; 926 private long nonce_ ; 927 /** 928 * <code>optional uint64 nonce = 4 [default = 0];</code> 929 */ hasNonce()930 public boolean hasNonce() { 931 return ((bitField0_ & 0x00000008) == 0x00000008); 932 } 933 /** 934 * <code>optional uint64 nonce = 4 [default = 0];</code> 935 */ getNonce()936 public long getNonce() { 937 return nonce_; 938 } 939 /** 940 * <code>optional uint64 nonce = 4 [default = 0];</code> 941 */ setNonce(long value)942 public Builder setNonce(long value) { 943 bitField0_ |= 0x00000008; 944 nonce_ = value; 945 onChanged(); 946 return this; 947 } 948 /** 949 * <code>optional uint64 nonce = 4 [default = 0];</code> 950 */ clearNonce()951 public Builder clearNonce() { 952 bitField0_ = (bitField0_ & ~0x00000008); 953 nonce_ = 0L; 954 onChanged(); 955 return this; 956 } 957 958 // @@protoc_insertion_point(builder_scope:AddColumnRequest) 959 } 960 961 static { 962 defaultInstance = new AddColumnRequest(true); defaultInstance.initFields()963 defaultInstance.initFields(); 964 } 965 966 // @@protoc_insertion_point(class_scope:AddColumnRequest) 967 } 968 969 public interface AddColumnResponseOrBuilder 970 extends com.google.protobuf.MessageOrBuilder { 971 } 972 /** 973 * Protobuf type {@code AddColumnResponse} 974 */ 975 public static final class AddColumnResponse extends 976 com.google.protobuf.GeneratedMessage 977 implements AddColumnResponseOrBuilder { 978 // Use AddColumnResponse.newBuilder() to construct. AddColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)979 private AddColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 980 super(builder); 981 this.unknownFields = builder.getUnknownFields(); 982 } AddColumnResponse(boolean noInit)983 private AddColumnResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 984 985 private static final AddColumnResponse defaultInstance; getDefaultInstance()986 public static AddColumnResponse getDefaultInstance() { 987 return defaultInstance; 988 } 989 getDefaultInstanceForType()990 public AddColumnResponse getDefaultInstanceForType() { 991 return defaultInstance; 992 } 993 994 private final com.google.protobuf.UnknownFieldSet unknownFields; 995 @java.lang.Override 996 public final com.google.protobuf.UnknownFieldSet getUnknownFields()997 getUnknownFields() { 998 return this.unknownFields; 999 } AddColumnResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1000 private AddColumnResponse( 1001 com.google.protobuf.CodedInputStream input, 1002 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1003 throws com.google.protobuf.InvalidProtocolBufferException { 1004 initFields(); 1005 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1006 com.google.protobuf.UnknownFieldSet.newBuilder(); 1007 try { 1008 boolean done = false; 1009 while (!done) { 1010 int tag = input.readTag(); 1011 switch (tag) { 1012 case 0: 1013 done = true; 1014 break; 1015 default: { 1016 if (!parseUnknownField(input, unknownFields, 1017 extensionRegistry, tag)) { 1018 done = true; 1019 } 1020 break; 1021 } 1022 } 1023 } 1024 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1025 throw e.setUnfinishedMessage(this); 1026 } catch (java.io.IOException e) { 1027 throw new com.google.protobuf.InvalidProtocolBufferException( 1028 e.getMessage()).setUnfinishedMessage(this); 1029 } finally { 1030 this.unknownFields = unknownFields.build(); 1031 makeExtensionsImmutable(); 1032 } 1033 } 1034 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()1035 getDescriptor() { 1036 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnResponse_descriptor; 1037 } 1038 1039 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()1040 internalGetFieldAccessorTable() { 1041 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnResponse_fieldAccessorTable 1042 .ensureFieldAccessorsInitialized( 1043 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.Builder.class); 1044 } 1045 1046 public static com.google.protobuf.Parser<AddColumnResponse> PARSER = 1047 new com.google.protobuf.AbstractParser<AddColumnResponse>() { 1048 public AddColumnResponse parsePartialFrom( 1049 com.google.protobuf.CodedInputStream input, 1050 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1051 throws com.google.protobuf.InvalidProtocolBufferException { 1052 return new AddColumnResponse(input, extensionRegistry); 1053 } 1054 }; 1055 1056 @java.lang.Override getParserForType()1057 public com.google.protobuf.Parser<AddColumnResponse> getParserForType() { 1058 return PARSER; 1059 } 1060 initFields()1061 private void initFields() { 1062 } 1063 private byte memoizedIsInitialized = -1; isInitialized()1064 public final boolean isInitialized() { 1065 byte isInitialized = memoizedIsInitialized; 1066 if (isInitialized != -1) return isInitialized == 1; 1067 1068 memoizedIsInitialized = 1; 1069 return true; 1070 } 1071 writeTo(com.google.protobuf.CodedOutputStream output)1072 public void writeTo(com.google.protobuf.CodedOutputStream output) 1073 throws java.io.IOException { 1074 getSerializedSize(); 1075 getUnknownFields().writeTo(output); 1076 } 1077 1078 private int memoizedSerializedSize = -1; getSerializedSize()1079 public int getSerializedSize() { 1080 int size = memoizedSerializedSize; 1081 if (size != -1) return size; 1082 1083 size = 0; 1084 size += getUnknownFields().getSerializedSize(); 1085 memoizedSerializedSize = size; 1086 return size; 1087 } 1088 1089 private static final long serialVersionUID = 0L; 1090 @java.lang.Override writeReplace()1091 protected java.lang.Object writeReplace() 1092 throws java.io.ObjectStreamException { 1093 return super.writeReplace(); 1094 } 1095 1096 @java.lang.Override equals(final java.lang.Object obj)1097 public boolean equals(final java.lang.Object obj) { 1098 if (obj == this) { 1099 return true; 1100 } 1101 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse)) { 1102 return super.equals(obj); 1103 } 1104 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse) obj; 1105 1106 boolean result = true; 1107 result = result && 1108 getUnknownFields().equals(other.getUnknownFields()); 1109 return result; 1110 } 1111 1112 private int memoizedHashCode = 0; 1113 @java.lang.Override hashCode()1114 public int hashCode() { 1115 if (memoizedHashCode != 0) { 1116 return memoizedHashCode; 1117 } 1118 int hash = 41; 1119 hash = (19 * hash) + getDescriptorForType().hashCode(); 1120 hash = (29 * hash) + getUnknownFields().hashCode(); 1121 memoizedHashCode = hash; 1122 return hash; 1123 } 1124 parseFrom( com.google.protobuf.ByteString data)1125 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1126 com.google.protobuf.ByteString data) 1127 throws com.google.protobuf.InvalidProtocolBufferException { 1128 return PARSER.parseFrom(data); 1129 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1130 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1131 com.google.protobuf.ByteString data, 1132 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1133 throws com.google.protobuf.InvalidProtocolBufferException { 1134 return PARSER.parseFrom(data, extensionRegistry); 1135 } parseFrom(byte[] data)1136 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(byte[] data) 1137 throws com.google.protobuf.InvalidProtocolBufferException { 1138 return PARSER.parseFrom(data); 1139 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1140 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1141 byte[] data, 1142 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1143 throws com.google.protobuf.InvalidProtocolBufferException { 1144 return PARSER.parseFrom(data, extensionRegistry); 1145 } parseFrom(java.io.InputStream input)1146 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(java.io.InputStream input) 1147 throws java.io.IOException { 1148 return PARSER.parseFrom(input); 1149 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1150 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1151 java.io.InputStream input, 1152 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1153 throws java.io.IOException { 1154 return PARSER.parseFrom(input, extensionRegistry); 1155 } parseDelimitedFrom(java.io.InputStream input)1156 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom(java.io.InputStream input) 1157 throws java.io.IOException { 1158 return PARSER.parseDelimitedFrom(input); 1159 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1160 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom( 1161 java.io.InputStream input, 1162 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1163 throws java.io.IOException { 1164 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1165 } parseFrom( com.google.protobuf.CodedInputStream input)1166 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1167 com.google.protobuf.CodedInputStream input) 1168 throws java.io.IOException { 1169 return PARSER.parseFrom(input); 1170 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1171 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parseFrom( 1172 com.google.protobuf.CodedInputStream input, 1173 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1174 throws java.io.IOException { 1175 return PARSER.parseFrom(input, extensionRegistry); 1176 } 1177 newBuilder()1178 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()1179 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse prototype)1180 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse prototype) { 1181 return newBuilder().mergeFrom(prototype); 1182 } toBuilder()1183 public Builder toBuilder() { return newBuilder(this); } 1184 1185 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)1186 protected Builder newBuilderForType( 1187 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1188 Builder builder = new Builder(parent); 1189 return builder; 1190 } 1191 /** 1192 * Protobuf type {@code AddColumnResponse} 1193 */ 1194 public static final class Builder extends 1195 com.google.protobuf.GeneratedMessage.Builder<Builder> 1196 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponseOrBuilder { 1197 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()1198 getDescriptor() { 1199 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnResponse_descriptor; 1200 } 1201 1202 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()1203 internalGetFieldAccessorTable() { 1204 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnResponse_fieldAccessorTable 1205 .ensureFieldAccessorsInitialized( 1206 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.Builder.class); 1207 } 1208 1209 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.newBuilder() Builder()1210 private Builder() { 1211 maybeForceBuilderInitialization(); 1212 } 1213 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)1214 private Builder( 1215 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1216 super(parent); 1217 maybeForceBuilderInitialization(); 1218 } maybeForceBuilderInitialization()1219 private void maybeForceBuilderInitialization() { 1220 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1221 } 1222 } create()1223 private static Builder create() { 1224 return new Builder(); 1225 } 1226 clear()1227 public Builder clear() { 1228 super.clear(); 1229 return this; 1230 } 1231 clone()1232 public Builder clone() { 1233 return create().mergeFrom(buildPartial()); 1234 } 1235 1236 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()1237 getDescriptorForType() { 1238 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AddColumnResponse_descriptor; 1239 } 1240 getDefaultInstanceForType()1241 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse getDefaultInstanceForType() { 1242 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance(); 1243 } 1244 build()1245 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse build() { 1246 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse result = buildPartial(); 1247 if (!result.isInitialized()) { 1248 throw newUninitializedMessageException(result); 1249 } 1250 return result; 1251 } 1252 buildPartial()1253 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse buildPartial() { 1254 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse(this); 1255 onBuilt(); 1256 return result; 1257 } 1258 mergeFrom(com.google.protobuf.Message other)1259 public Builder mergeFrom(com.google.protobuf.Message other) { 1260 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse) { 1261 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse)other); 1262 } else { 1263 super.mergeFrom(other); 1264 return this; 1265 } 1266 } 1267 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse other)1268 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse other) { 1269 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance()) return this; 1270 this.mergeUnknownFields(other.getUnknownFields()); 1271 return this; 1272 } 1273 isInitialized()1274 public final boolean isInitialized() { 1275 return true; 1276 } 1277 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1278 public Builder mergeFrom( 1279 com.google.protobuf.CodedInputStream input, 1280 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1281 throws java.io.IOException { 1282 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse parsedMessage = null; 1283 try { 1284 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1285 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1286 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AddColumnResponse) e.getUnfinishedMessage(); 1287 throw e; 1288 } finally { 1289 if (parsedMessage != null) { 1290 mergeFrom(parsedMessage); 1291 } 1292 } 1293 return this; 1294 } 1295 1296 // @@protoc_insertion_point(builder_scope:AddColumnResponse) 1297 } 1298 1299 static { 1300 defaultInstance = new AddColumnResponse(true); defaultInstance.initFields()1301 defaultInstance.initFields(); 1302 } 1303 1304 // @@protoc_insertion_point(class_scope:AddColumnResponse) 1305 } 1306 1307 public interface DeleteColumnRequestOrBuilder 1308 extends com.google.protobuf.MessageOrBuilder { 1309 1310 // required .TableName table_name = 1; 1311 /** 1312 * <code>required .TableName table_name = 1;</code> 1313 */ hasTableName()1314 boolean hasTableName(); 1315 /** 1316 * <code>required .TableName table_name = 1;</code> 1317 */ getTableName()1318 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); 1319 /** 1320 * <code>required .TableName table_name = 1;</code> 1321 */ getTableNameOrBuilder()1322 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); 1323 1324 // required bytes column_name = 2; 1325 /** 1326 * <code>required bytes column_name = 2;</code> 1327 */ hasColumnName()1328 boolean hasColumnName(); 1329 /** 1330 * <code>required bytes column_name = 2;</code> 1331 */ getColumnName()1332 com.google.protobuf.ByteString getColumnName(); 1333 1334 // optional uint64 nonce_group = 3 [default = 0]; 1335 /** 1336 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 1337 */ hasNonceGroup()1338 boolean hasNonceGroup(); 1339 /** 1340 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 1341 */ getNonceGroup()1342 long getNonceGroup(); 1343 1344 // optional uint64 nonce = 4 [default = 0]; 1345 /** 1346 * <code>optional uint64 nonce = 4 [default = 0];</code> 1347 */ hasNonce()1348 boolean hasNonce(); 1349 /** 1350 * <code>optional uint64 nonce = 4 [default = 0];</code> 1351 */ getNonce()1352 long getNonce(); 1353 } 1354 /** 1355 * Protobuf type {@code DeleteColumnRequest} 1356 */ 1357 public static final class DeleteColumnRequest extends 1358 com.google.protobuf.GeneratedMessage 1359 implements DeleteColumnRequestOrBuilder { 1360 // Use DeleteColumnRequest.newBuilder() to construct. DeleteColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)1361 private DeleteColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1362 super(builder); 1363 this.unknownFields = builder.getUnknownFields(); 1364 } DeleteColumnRequest(boolean noInit)1365 private DeleteColumnRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1366 1367 private static final DeleteColumnRequest defaultInstance; getDefaultInstance()1368 public static DeleteColumnRequest getDefaultInstance() { 1369 return defaultInstance; 1370 } 1371 getDefaultInstanceForType()1372 public DeleteColumnRequest getDefaultInstanceForType() { 1373 return defaultInstance; 1374 } 1375 1376 private final com.google.protobuf.UnknownFieldSet unknownFields; 1377 @java.lang.Override 1378 public final com.google.protobuf.UnknownFieldSet getUnknownFields()1379 getUnknownFields() { 1380 return this.unknownFields; 1381 } DeleteColumnRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1382 private DeleteColumnRequest( 1383 com.google.protobuf.CodedInputStream input, 1384 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1385 throws com.google.protobuf.InvalidProtocolBufferException { 1386 initFields(); 1387 int mutable_bitField0_ = 0; 1388 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1389 com.google.protobuf.UnknownFieldSet.newBuilder(); 1390 try { 1391 boolean done = false; 1392 while (!done) { 1393 int tag = input.readTag(); 1394 switch (tag) { 1395 case 0: 1396 done = true; 1397 break; 1398 default: { 1399 if (!parseUnknownField(input, unknownFields, 1400 extensionRegistry, tag)) { 1401 done = true; 1402 } 1403 break; 1404 } 1405 case 10: { 1406 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; 1407 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1408 subBuilder = tableName_.toBuilder(); 1409 } 1410 tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); 1411 if (subBuilder != null) { 1412 subBuilder.mergeFrom(tableName_); 1413 tableName_ = subBuilder.buildPartial(); 1414 } 1415 bitField0_ |= 0x00000001; 1416 break; 1417 } 1418 case 18: { 1419 bitField0_ |= 0x00000002; 1420 columnName_ = input.readBytes(); 1421 break; 1422 } 1423 case 24: { 1424 bitField0_ |= 0x00000004; 1425 nonceGroup_ = input.readUInt64(); 1426 break; 1427 } 1428 case 32: { 1429 bitField0_ |= 0x00000008; 1430 nonce_ = input.readUInt64(); 1431 break; 1432 } 1433 } 1434 } 1435 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1436 throw e.setUnfinishedMessage(this); 1437 } catch (java.io.IOException e) { 1438 throw new com.google.protobuf.InvalidProtocolBufferException( 1439 e.getMessage()).setUnfinishedMessage(this); 1440 } finally { 1441 this.unknownFields = unknownFields.build(); 1442 makeExtensionsImmutable(); 1443 } 1444 } 1445 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()1446 getDescriptor() { 1447 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnRequest_descriptor; 1448 } 1449 1450 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()1451 internalGetFieldAccessorTable() { 1452 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnRequest_fieldAccessorTable 1453 .ensureFieldAccessorsInitialized( 1454 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.Builder.class); 1455 } 1456 1457 public static com.google.protobuf.Parser<DeleteColumnRequest> PARSER = 1458 new com.google.protobuf.AbstractParser<DeleteColumnRequest>() { 1459 public DeleteColumnRequest parsePartialFrom( 1460 com.google.protobuf.CodedInputStream input, 1461 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1462 throws com.google.protobuf.InvalidProtocolBufferException { 1463 return new DeleteColumnRequest(input, extensionRegistry); 1464 } 1465 }; 1466 1467 @java.lang.Override getParserForType()1468 public com.google.protobuf.Parser<DeleteColumnRequest> getParserForType() { 1469 return PARSER; 1470 } 1471 1472 private int bitField0_; 1473 // required .TableName table_name = 1; 1474 public static final int TABLE_NAME_FIELD_NUMBER = 1; 1475 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; 1476 /** 1477 * <code>required .TableName table_name = 1;</code> 1478 */ hasTableName()1479 public boolean hasTableName() { 1480 return ((bitField0_ & 0x00000001) == 0x00000001); 1481 } 1482 /** 1483 * <code>required .TableName table_name = 1;</code> 1484 */ getTableName()1485 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 1486 return tableName_; 1487 } 1488 /** 1489 * <code>required .TableName table_name = 1;</code> 1490 */ getTableNameOrBuilder()1491 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 1492 return tableName_; 1493 } 1494 1495 // required bytes column_name = 2; 1496 public static final int COLUMN_NAME_FIELD_NUMBER = 2; 1497 private com.google.protobuf.ByteString columnName_; 1498 /** 1499 * <code>required bytes column_name = 2;</code> 1500 */ hasColumnName()1501 public boolean hasColumnName() { 1502 return ((bitField0_ & 0x00000002) == 0x00000002); 1503 } 1504 /** 1505 * <code>required bytes column_name = 2;</code> 1506 */ getColumnName()1507 public com.google.protobuf.ByteString getColumnName() { 1508 return columnName_; 1509 } 1510 1511 // optional uint64 nonce_group = 3 [default = 0]; 1512 public static final int NONCE_GROUP_FIELD_NUMBER = 3; 1513 private long nonceGroup_; 1514 /** 1515 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 1516 */ hasNonceGroup()1517 public boolean hasNonceGroup() { 1518 return ((bitField0_ & 0x00000004) == 0x00000004); 1519 } 1520 /** 1521 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 1522 */ getNonceGroup()1523 public long getNonceGroup() { 1524 return nonceGroup_; 1525 } 1526 1527 // optional uint64 nonce = 4 [default = 0]; 1528 public static final int NONCE_FIELD_NUMBER = 4; 1529 private long nonce_; 1530 /** 1531 * <code>optional uint64 nonce = 4 [default = 0];</code> 1532 */ hasNonce()1533 public boolean hasNonce() { 1534 return ((bitField0_ & 0x00000008) == 0x00000008); 1535 } 1536 /** 1537 * <code>optional uint64 nonce = 4 [default = 0];</code> 1538 */ getNonce()1539 public long getNonce() { 1540 return nonce_; 1541 } 1542 initFields()1543 private void initFields() { 1544 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 1545 columnName_ = com.google.protobuf.ByteString.EMPTY; 1546 nonceGroup_ = 0L; 1547 nonce_ = 0L; 1548 } 1549 private byte memoizedIsInitialized = -1; isInitialized()1550 public final boolean isInitialized() { 1551 byte isInitialized = memoizedIsInitialized; 1552 if (isInitialized != -1) return isInitialized == 1; 1553 1554 if (!hasTableName()) { 1555 memoizedIsInitialized = 0; 1556 return false; 1557 } 1558 if (!hasColumnName()) { 1559 memoizedIsInitialized = 0; 1560 return false; 1561 } 1562 if (!getTableName().isInitialized()) { 1563 memoizedIsInitialized = 0; 1564 return false; 1565 } 1566 memoizedIsInitialized = 1; 1567 return true; 1568 } 1569 writeTo(com.google.protobuf.CodedOutputStream output)1570 public void writeTo(com.google.protobuf.CodedOutputStream output) 1571 throws java.io.IOException { 1572 getSerializedSize(); 1573 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1574 output.writeMessage(1, tableName_); 1575 } 1576 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1577 output.writeBytes(2, columnName_); 1578 } 1579 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1580 output.writeUInt64(3, nonceGroup_); 1581 } 1582 if (((bitField0_ & 0x00000008) == 0x00000008)) { 1583 output.writeUInt64(4, nonce_); 1584 } 1585 getUnknownFields().writeTo(output); 1586 } 1587 1588 private int memoizedSerializedSize = -1; getSerializedSize()1589 public int getSerializedSize() { 1590 int size = memoizedSerializedSize; 1591 if (size != -1) return size; 1592 1593 size = 0; 1594 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1595 size += com.google.protobuf.CodedOutputStream 1596 .computeMessageSize(1, tableName_); 1597 } 1598 if (((bitField0_ & 0x00000002) == 0x00000002)) { 1599 size += com.google.protobuf.CodedOutputStream 1600 .computeBytesSize(2, columnName_); 1601 } 1602 if (((bitField0_ & 0x00000004) == 0x00000004)) { 1603 size += com.google.protobuf.CodedOutputStream 1604 .computeUInt64Size(3, nonceGroup_); 1605 } 1606 if (((bitField0_ & 0x00000008) == 0x00000008)) { 1607 size += com.google.protobuf.CodedOutputStream 1608 .computeUInt64Size(4, nonce_); 1609 } 1610 size += getUnknownFields().getSerializedSize(); 1611 memoizedSerializedSize = size; 1612 return size; 1613 } 1614 1615 private static final long serialVersionUID = 0L; 1616 @java.lang.Override writeReplace()1617 protected java.lang.Object writeReplace() 1618 throws java.io.ObjectStreamException { 1619 return super.writeReplace(); 1620 } 1621 1622 @java.lang.Override equals(final java.lang.Object obj)1623 public boolean equals(final java.lang.Object obj) { 1624 if (obj == this) { 1625 return true; 1626 } 1627 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest)) { 1628 return super.equals(obj); 1629 } 1630 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest) obj; 1631 1632 boolean result = true; 1633 result = result && (hasTableName() == other.hasTableName()); 1634 if (hasTableName()) { 1635 result = result && getTableName() 1636 .equals(other.getTableName()); 1637 } 1638 result = result && (hasColumnName() == other.hasColumnName()); 1639 if (hasColumnName()) { 1640 result = result && getColumnName() 1641 .equals(other.getColumnName()); 1642 } 1643 result = result && (hasNonceGroup() == other.hasNonceGroup()); 1644 if (hasNonceGroup()) { 1645 result = result && (getNonceGroup() 1646 == other.getNonceGroup()); 1647 } 1648 result = result && (hasNonce() == other.hasNonce()); 1649 if (hasNonce()) { 1650 result = result && (getNonce() 1651 == other.getNonce()); 1652 } 1653 result = result && 1654 getUnknownFields().equals(other.getUnknownFields()); 1655 return result; 1656 } 1657 1658 private int memoizedHashCode = 0; 1659 @java.lang.Override hashCode()1660 public int hashCode() { 1661 if (memoizedHashCode != 0) { 1662 return memoizedHashCode; 1663 } 1664 int hash = 41; 1665 hash = (19 * hash) + getDescriptorForType().hashCode(); 1666 if (hasTableName()) { 1667 hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; 1668 hash = (53 * hash) + getTableName().hashCode(); 1669 } 1670 if (hasColumnName()) { 1671 hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER; 1672 hash = (53 * hash) + getColumnName().hashCode(); 1673 } 1674 if (hasNonceGroup()) { 1675 hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; 1676 hash = (53 * hash) + hashLong(getNonceGroup()); 1677 } 1678 if (hasNonce()) { 1679 hash = (37 * hash) + NONCE_FIELD_NUMBER; 1680 hash = (53 * hash) + hashLong(getNonce()); 1681 } 1682 hash = (29 * hash) + getUnknownFields().hashCode(); 1683 memoizedHashCode = hash; 1684 return hash; 1685 } 1686 parseFrom( com.google.protobuf.ByteString data)1687 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1688 com.google.protobuf.ByteString data) 1689 throws com.google.protobuf.InvalidProtocolBufferException { 1690 return PARSER.parseFrom(data); 1691 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1692 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1693 com.google.protobuf.ByteString data, 1694 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1695 throws com.google.protobuf.InvalidProtocolBufferException { 1696 return PARSER.parseFrom(data, extensionRegistry); 1697 } parseFrom(byte[] data)1698 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(byte[] data) 1699 throws com.google.protobuf.InvalidProtocolBufferException { 1700 return PARSER.parseFrom(data); 1701 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1702 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1703 byte[] data, 1704 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1705 throws com.google.protobuf.InvalidProtocolBufferException { 1706 return PARSER.parseFrom(data, extensionRegistry); 1707 } parseFrom(java.io.InputStream input)1708 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(java.io.InputStream input) 1709 throws java.io.IOException { 1710 return PARSER.parseFrom(input); 1711 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1712 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1713 java.io.InputStream input, 1714 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1715 throws java.io.IOException { 1716 return PARSER.parseFrom(input, extensionRegistry); 1717 } parseDelimitedFrom(java.io.InputStream input)1718 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom(java.io.InputStream input) 1719 throws java.io.IOException { 1720 return PARSER.parseDelimitedFrom(input); 1721 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1722 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom( 1723 java.io.InputStream input, 1724 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1725 throws java.io.IOException { 1726 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1727 } parseFrom( com.google.protobuf.CodedInputStream input)1728 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1729 com.google.protobuf.CodedInputStream input) 1730 throws java.io.IOException { 1731 return PARSER.parseFrom(input); 1732 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1733 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom( 1734 com.google.protobuf.CodedInputStream input, 1735 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1736 throws java.io.IOException { 1737 return PARSER.parseFrom(input, extensionRegistry); 1738 } 1739 newBuilder()1740 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()1741 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest prototype)1742 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest prototype) { 1743 return newBuilder().mergeFrom(prototype); 1744 } toBuilder()1745 public Builder toBuilder() { return newBuilder(this); } 1746 1747 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)1748 protected Builder newBuilderForType( 1749 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1750 Builder builder = new Builder(parent); 1751 return builder; 1752 } 1753 /** 1754 * Protobuf type {@code DeleteColumnRequest} 1755 */ 1756 public static final class Builder extends 1757 com.google.protobuf.GeneratedMessage.Builder<Builder> 1758 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequestOrBuilder { 1759 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()1760 getDescriptor() { 1761 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnRequest_descriptor; 1762 } 1763 1764 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()1765 internalGetFieldAccessorTable() { 1766 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnRequest_fieldAccessorTable 1767 .ensureFieldAccessorsInitialized( 1768 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.Builder.class); 1769 } 1770 1771 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.newBuilder() Builder()1772 private Builder() { 1773 maybeForceBuilderInitialization(); 1774 } 1775 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)1776 private Builder( 1777 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1778 super(parent); 1779 maybeForceBuilderInitialization(); 1780 } maybeForceBuilderInitialization()1781 private void maybeForceBuilderInitialization() { 1782 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1783 getTableNameFieldBuilder(); 1784 } 1785 } create()1786 private static Builder create() { 1787 return new Builder(); 1788 } 1789 clear()1790 public Builder clear() { 1791 super.clear(); 1792 if (tableNameBuilder_ == null) { 1793 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 1794 } else { 1795 tableNameBuilder_.clear(); 1796 } 1797 bitField0_ = (bitField0_ & ~0x00000001); 1798 columnName_ = com.google.protobuf.ByteString.EMPTY; 1799 bitField0_ = (bitField0_ & ~0x00000002); 1800 nonceGroup_ = 0L; 1801 bitField0_ = (bitField0_ & ~0x00000004); 1802 nonce_ = 0L; 1803 bitField0_ = (bitField0_ & ~0x00000008); 1804 return this; 1805 } 1806 clone()1807 public Builder clone() { 1808 return create().mergeFrom(buildPartial()); 1809 } 1810 1811 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()1812 getDescriptorForType() { 1813 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnRequest_descriptor; 1814 } 1815 getDefaultInstanceForType()1816 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest getDefaultInstanceForType() { 1817 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance(); 1818 } 1819 build()1820 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest build() { 1821 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest result = buildPartial(); 1822 if (!result.isInitialized()) { 1823 throw newUninitializedMessageException(result); 1824 } 1825 return result; 1826 } 1827 buildPartial()1828 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest buildPartial() { 1829 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest(this); 1830 int from_bitField0_ = bitField0_; 1831 int to_bitField0_ = 0; 1832 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1833 to_bitField0_ |= 0x00000001; 1834 } 1835 if (tableNameBuilder_ == null) { 1836 result.tableName_ = tableName_; 1837 } else { 1838 result.tableName_ = tableNameBuilder_.build(); 1839 } 1840 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 1841 to_bitField0_ |= 0x00000002; 1842 } 1843 result.columnName_ = columnName_; 1844 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 1845 to_bitField0_ |= 0x00000004; 1846 } 1847 result.nonceGroup_ = nonceGroup_; 1848 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 1849 to_bitField0_ |= 0x00000008; 1850 } 1851 result.nonce_ = nonce_; 1852 result.bitField0_ = to_bitField0_; 1853 onBuilt(); 1854 return result; 1855 } 1856 mergeFrom(com.google.protobuf.Message other)1857 public Builder mergeFrom(com.google.protobuf.Message other) { 1858 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest) { 1859 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest)other); 1860 } else { 1861 super.mergeFrom(other); 1862 return this; 1863 } 1864 } 1865 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest other)1866 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest other) { 1867 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance()) return this; 1868 if (other.hasTableName()) { 1869 mergeTableName(other.getTableName()); 1870 } 1871 if (other.hasColumnName()) { 1872 setColumnName(other.getColumnName()); 1873 } 1874 if (other.hasNonceGroup()) { 1875 setNonceGroup(other.getNonceGroup()); 1876 } 1877 if (other.hasNonce()) { 1878 setNonce(other.getNonce()); 1879 } 1880 this.mergeUnknownFields(other.getUnknownFields()); 1881 return this; 1882 } 1883 isInitialized()1884 public final boolean isInitialized() { 1885 if (!hasTableName()) { 1886 1887 return false; 1888 } 1889 if (!hasColumnName()) { 1890 1891 return false; 1892 } 1893 if (!getTableName().isInitialized()) { 1894 1895 return false; 1896 } 1897 return true; 1898 } 1899 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)1900 public Builder mergeFrom( 1901 com.google.protobuf.CodedInputStream input, 1902 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1903 throws java.io.IOException { 1904 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest parsedMessage = null; 1905 try { 1906 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1907 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1908 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnRequest) e.getUnfinishedMessage(); 1909 throw e; 1910 } finally { 1911 if (parsedMessage != null) { 1912 mergeFrom(parsedMessage); 1913 } 1914 } 1915 return this; 1916 } 1917 private int bitField0_; 1918 1919 // required .TableName table_name = 1; 1920 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 1921 private com.google.protobuf.SingleFieldBuilder< 1922 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; 1923 /** 1924 * <code>required .TableName table_name = 1;</code> 1925 */ hasTableName()1926 public boolean hasTableName() { 1927 return ((bitField0_ & 0x00000001) == 0x00000001); 1928 } 1929 /** 1930 * <code>required .TableName table_name = 1;</code> 1931 */ getTableName()1932 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 1933 if (tableNameBuilder_ == null) { 1934 return tableName_; 1935 } else { 1936 return tableNameBuilder_.getMessage(); 1937 } 1938 } 1939 /** 1940 * <code>required .TableName table_name = 1;</code> 1941 */ setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)1942 public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 1943 if (tableNameBuilder_ == null) { 1944 if (value == null) { 1945 throw new NullPointerException(); 1946 } 1947 tableName_ = value; 1948 onChanged(); 1949 } else { 1950 tableNameBuilder_.setMessage(value); 1951 } 1952 bitField0_ |= 0x00000001; 1953 return this; 1954 } 1955 /** 1956 * <code>required .TableName table_name = 1;</code> 1957 */ setTableName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue)1958 public Builder setTableName( 1959 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { 1960 if (tableNameBuilder_ == null) { 1961 tableName_ = builderForValue.build(); 1962 onChanged(); 1963 } else { 1964 tableNameBuilder_.setMessage(builderForValue.build()); 1965 } 1966 bitField0_ |= 0x00000001; 1967 return this; 1968 } 1969 /** 1970 * <code>required .TableName table_name = 1;</code> 1971 */ mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)1972 public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 1973 if (tableNameBuilder_ == null) { 1974 if (((bitField0_ & 0x00000001) == 0x00000001) && 1975 tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { 1976 tableName_ = 1977 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); 1978 } else { 1979 tableName_ = value; 1980 } 1981 onChanged(); 1982 } else { 1983 tableNameBuilder_.mergeFrom(value); 1984 } 1985 bitField0_ |= 0x00000001; 1986 return this; 1987 } 1988 /** 1989 * <code>required .TableName table_name = 1;</code> 1990 */ clearTableName()1991 public Builder clearTableName() { 1992 if (tableNameBuilder_ == null) { 1993 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 1994 onChanged(); 1995 } else { 1996 tableNameBuilder_.clear(); 1997 } 1998 bitField0_ = (bitField0_ & ~0x00000001); 1999 return this; 2000 } 2001 /** 2002 * <code>required .TableName table_name = 1;</code> 2003 */ getTableNameBuilder()2004 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { 2005 bitField0_ |= 0x00000001; 2006 onChanged(); 2007 return getTableNameFieldBuilder().getBuilder(); 2008 } 2009 /** 2010 * <code>required .TableName table_name = 1;</code> 2011 */ getTableNameOrBuilder()2012 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 2013 if (tableNameBuilder_ != null) { 2014 return tableNameBuilder_.getMessageOrBuilder(); 2015 } else { 2016 return tableName_; 2017 } 2018 } 2019 /** 2020 * <code>required .TableName table_name = 1;</code> 2021 */ 2022 private com.google.protobuf.SingleFieldBuilder< 2023 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder()2024 getTableNameFieldBuilder() { 2025 if (tableNameBuilder_ == null) { 2026 tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< 2027 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( 2028 tableName_, 2029 getParentForChildren(), 2030 isClean()); 2031 tableName_ = null; 2032 } 2033 return tableNameBuilder_; 2034 } 2035 2036 // required bytes column_name = 2; 2037 private com.google.protobuf.ByteString columnName_ = com.google.protobuf.ByteString.EMPTY; 2038 /** 2039 * <code>required bytes column_name = 2;</code> 2040 */ hasColumnName()2041 public boolean hasColumnName() { 2042 return ((bitField0_ & 0x00000002) == 0x00000002); 2043 } 2044 /** 2045 * <code>required bytes column_name = 2;</code> 2046 */ getColumnName()2047 public com.google.protobuf.ByteString getColumnName() { 2048 return columnName_; 2049 } 2050 /** 2051 * <code>required bytes column_name = 2;</code> 2052 */ setColumnName(com.google.protobuf.ByteString value)2053 public Builder setColumnName(com.google.protobuf.ByteString value) { 2054 if (value == null) { 2055 throw new NullPointerException(); 2056 } 2057 bitField0_ |= 0x00000002; 2058 columnName_ = value; 2059 onChanged(); 2060 return this; 2061 } 2062 /** 2063 * <code>required bytes column_name = 2;</code> 2064 */ clearColumnName()2065 public Builder clearColumnName() { 2066 bitField0_ = (bitField0_ & ~0x00000002); 2067 columnName_ = getDefaultInstance().getColumnName(); 2068 onChanged(); 2069 return this; 2070 } 2071 2072 // optional uint64 nonce_group = 3 [default = 0]; 2073 private long nonceGroup_ ; 2074 /** 2075 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2076 */ hasNonceGroup()2077 public boolean hasNonceGroup() { 2078 return ((bitField0_ & 0x00000004) == 0x00000004); 2079 } 2080 /** 2081 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2082 */ getNonceGroup()2083 public long getNonceGroup() { 2084 return nonceGroup_; 2085 } 2086 /** 2087 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2088 */ setNonceGroup(long value)2089 public Builder setNonceGroup(long value) { 2090 bitField0_ |= 0x00000004; 2091 nonceGroup_ = value; 2092 onChanged(); 2093 return this; 2094 } 2095 /** 2096 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2097 */ clearNonceGroup()2098 public Builder clearNonceGroup() { 2099 bitField0_ = (bitField0_ & ~0x00000004); 2100 nonceGroup_ = 0L; 2101 onChanged(); 2102 return this; 2103 } 2104 2105 // optional uint64 nonce = 4 [default = 0]; 2106 private long nonce_ ; 2107 /** 2108 * <code>optional uint64 nonce = 4 [default = 0];</code> 2109 */ hasNonce()2110 public boolean hasNonce() { 2111 return ((bitField0_ & 0x00000008) == 0x00000008); 2112 } 2113 /** 2114 * <code>optional uint64 nonce = 4 [default = 0];</code> 2115 */ getNonce()2116 public long getNonce() { 2117 return nonce_; 2118 } 2119 /** 2120 * <code>optional uint64 nonce = 4 [default = 0];</code> 2121 */ setNonce(long value)2122 public Builder setNonce(long value) { 2123 bitField0_ |= 0x00000008; 2124 nonce_ = value; 2125 onChanged(); 2126 return this; 2127 } 2128 /** 2129 * <code>optional uint64 nonce = 4 [default = 0];</code> 2130 */ clearNonce()2131 public Builder clearNonce() { 2132 bitField0_ = (bitField0_ & ~0x00000008); 2133 nonce_ = 0L; 2134 onChanged(); 2135 return this; 2136 } 2137 2138 // @@protoc_insertion_point(builder_scope:DeleteColumnRequest) 2139 } 2140 2141 static { 2142 defaultInstance = new DeleteColumnRequest(true); defaultInstance.initFields()2143 defaultInstance.initFields(); 2144 } 2145 2146 // @@protoc_insertion_point(class_scope:DeleteColumnRequest) 2147 } 2148 2149 public interface DeleteColumnResponseOrBuilder 2150 extends com.google.protobuf.MessageOrBuilder { 2151 } 2152 /** 2153 * Protobuf type {@code DeleteColumnResponse} 2154 */ 2155 public static final class DeleteColumnResponse extends 2156 com.google.protobuf.GeneratedMessage 2157 implements DeleteColumnResponseOrBuilder { 2158 // Use DeleteColumnResponse.newBuilder() to construct. DeleteColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)2159 private DeleteColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2160 super(builder); 2161 this.unknownFields = builder.getUnknownFields(); 2162 } DeleteColumnResponse(boolean noInit)2163 private DeleteColumnResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2164 2165 private static final DeleteColumnResponse defaultInstance; getDefaultInstance()2166 public static DeleteColumnResponse getDefaultInstance() { 2167 return defaultInstance; 2168 } 2169 getDefaultInstanceForType()2170 public DeleteColumnResponse getDefaultInstanceForType() { 2171 return defaultInstance; 2172 } 2173 2174 private final com.google.protobuf.UnknownFieldSet unknownFields; 2175 @java.lang.Override 2176 public final com.google.protobuf.UnknownFieldSet getUnknownFields()2177 getUnknownFields() { 2178 return this.unknownFields; 2179 } DeleteColumnResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2180 private DeleteColumnResponse( 2181 com.google.protobuf.CodedInputStream input, 2182 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2183 throws com.google.protobuf.InvalidProtocolBufferException { 2184 initFields(); 2185 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2186 com.google.protobuf.UnknownFieldSet.newBuilder(); 2187 try { 2188 boolean done = false; 2189 while (!done) { 2190 int tag = input.readTag(); 2191 switch (tag) { 2192 case 0: 2193 done = true; 2194 break; 2195 default: { 2196 if (!parseUnknownField(input, unknownFields, 2197 extensionRegistry, tag)) { 2198 done = true; 2199 } 2200 break; 2201 } 2202 } 2203 } 2204 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2205 throw e.setUnfinishedMessage(this); 2206 } catch (java.io.IOException e) { 2207 throw new com.google.protobuf.InvalidProtocolBufferException( 2208 e.getMessage()).setUnfinishedMessage(this); 2209 } finally { 2210 this.unknownFields = unknownFields.build(); 2211 makeExtensionsImmutable(); 2212 } 2213 } 2214 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()2215 getDescriptor() { 2216 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnResponse_descriptor; 2217 } 2218 2219 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()2220 internalGetFieldAccessorTable() { 2221 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnResponse_fieldAccessorTable 2222 .ensureFieldAccessorsInitialized( 2223 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.Builder.class); 2224 } 2225 2226 public static com.google.protobuf.Parser<DeleteColumnResponse> PARSER = 2227 new com.google.protobuf.AbstractParser<DeleteColumnResponse>() { 2228 public DeleteColumnResponse parsePartialFrom( 2229 com.google.protobuf.CodedInputStream input, 2230 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2231 throws com.google.protobuf.InvalidProtocolBufferException { 2232 return new DeleteColumnResponse(input, extensionRegistry); 2233 } 2234 }; 2235 2236 @java.lang.Override getParserForType()2237 public com.google.protobuf.Parser<DeleteColumnResponse> getParserForType() { 2238 return PARSER; 2239 } 2240 initFields()2241 private void initFields() { 2242 } 2243 private byte memoizedIsInitialized = -1; isInitialized()2244 public final boolean isInitialized() { 2245 byte isInitialized = memoizedIsInitialized; 2246 if (isInitialized != -1) return isInitialized == 1; 2247 2248 memoizedIsInitialized = 1; 2249 return true; 2250 } 2251 writeTo(com.google.protobuf.CodedOutputStream output)2252 public void writeTo(com.google.protobuf.CodedOutputStream output) 2253 throws java.io.IOException { 2254 getSerializedSize(); 2255 getUnknownFields().writeTo(output); 2256 } 2257 2258 private int memoizedSerializedSize = -1; getSerializedSize()2259 public int getSerializedSize() { 2260 int size = memoizedSerializedSize; 2261 if (size != -1) return size; 2262 2263 size = 0; 2264 size += getUnknownFields().getSerializedSize(); 2265 memoizedSerializedSize = size; 2266 return size; 2267 } 2268 2269 private static final long serialVersionUID = 0L; 2270 @java.lang.Override writeReplace()2271 protected java.lang.Object writeReplace() 2272 throws java.io.ObjectStreamException { 2273 return super.writeReplace(); 2274 } 2275 2276 @java.lang.Override equals(final java.lang.Object obj)2277 public boolean equals(final java.lang.Object obj) { 2278 if (obj == this) { 2279 return true; 2280 } 2281 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse)) { 2282 return super.equals(obj); 2283 } 2284 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse) obj; 2285 2286 boolean result = true; 2287 result = result && 2288 getUnknownFields().equals(other.getUnknownFields()); 2289 return result; 2290 } 2291 2292 private int memoizedHashCode = 0; 2293 @java.lang.Override hashCode()2294 public int hashCode() { 2295 if (memoizedHashCode != 0) { 2296 return memoizedHashCode; 2297 } 2298 int hash = 41; 2299 hash = (19 * hash) + getDescriptorForType().hashCode(); 2300 hash = (29 * hash) + getUnknownFields().hashCode(); 2301 memoizedHashCode = hash; 2302 return hash; 2303 } 2304 parseFrom( com.google.protobuf.ByteString data)2305 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2306 com.google.protobuf.ByteString data) 2307 throws com.google.protobuf.InvalidProtocolBufferException { 2308 return PARSER.parseFrom(data); 2309 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2310 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2311 com.google.protobuf.ByteString data, 2312 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2313 throws com.google.protobuf.InvalidProtocolBufferException { 2314 return PARSER.parseFrom(data, extensionRegistry); 2315 } parseFrom(byte[] data)2316 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(byte[] data) 2317 throws com.google.protobuf.InvalidProtocolBufferException { 2318 return PARSER.parseFrom(data); 2319 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2320 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2321 byte[] data, 2322 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2323 throws com.google.protobuf.InvalidProtocolBufferException { 2324 return PARSER.parseFrom(data, extensionRegistry); 2325 } parseFrom(java.io.InputStream input)2326 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(java.io.InputStream input) 2327 throws java.io.IOException { 2328 return PARSER.parseFrom(input); 2329 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2330 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2331 java.io.InputStream input, 2332 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2333 throws java.io.IOException { 2334 return PARSER.parseFrom(input, extensionRegistry); 2335 } parseDelimitedFrom(java.io.InputStream input)2336 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom(java.io.InputStream input) 2337 throws java.io.IOException { 2338 return PARSER.parseDelimitedFrom(input); 2339 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2340 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom( 2341 java.io.InputStream input, 2342 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2343 throws java.io.IOException { 2344 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2345 } parseFrom( com.google.protobuf.CodedInputStream input)2346 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2347 com.google.protobuf.CodedInputStream input) 2348 throws java.io.IOException { 2349 return PARSER.parseFrom(input); 2350 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2351 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom( 2352 com.google.protobuf.CodedInputStream input, 2353 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2354 throws java.io.IOException { 2355 return PARSER.parseFrom(input, extensionRegistry); 2356 } 2357 newBuilder()2358 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()2359 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse prototype)2360 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse prototype) { 2361 return newBuilder().mergeFrom(prototype); 2362 } toBuilder()2363 public Builder toBuilder() { return newBuilder(this); } 2364 2365 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)2366 protected Builder newBuilderForType( 2367 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2368 Builder builder = new Builder(parent); 2369 return builder; 2370 } 2371 /** 2372 * Protobuf type {@code DeleteColumnResponse} 2373 */ 2374 public static final class Builder extends 2375 com.google.protobuf.GeneratedMessage.Builder<Builder> 2376 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponseOrBuilder { 2377 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()2378 getDescriptor() { 2379 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnResponse_descriptor; 2380 } 2381 2382 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()2383 internalGetFieldAccessorTable() { 2384 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnResponse_fieldAccessorTable 2385 .ensureFieldAccessorsInitialized( 2386 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.Builder.class); 2387 } 2388 2389 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.newBuilder() Builder()2390 private Builder() { 2391 maybeForceBuilderInitialization(); 2392 } 2393 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)2394 private Builder( 2395 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2396 super(parent); 2397 maybeForceBuilderInitialization(); 2398 } maybeForceBuilderInitialization()2399 private void maybeForceBuilderInitialization() { 2400 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2401 } 2402 } create()2403 private static Builder create() { 2404 return new Builder(); 2405 } 2406 clear()2407 public Builder clear() { 2408 super.clear(); 2409 return this; 2410 } 2411 clone()2412 public Builder clone() { 2413 return create().mergeFrom(buildPartial()); 2414 } 2415 2416 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()2417 getDescriptorForType() { 2418 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DeleteColumnResponse_descriptor; 2419 } 2420 getDefaultInstanceForType()2421 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse getDefaultInstanceForType() { 2422 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance(); 2423 } 2424 build()2425 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse build() { 2426 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse result = buildPartial(); 2427 if (!result.isInitialized()) { 2428 throw newUninitializedMessageException(result); 2429 } 2430 return result; 2431 } 2432 buildPartial()2433 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse buildPartial() { 2434 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse(this); 2435 onBuilt(); 2436 return result; 2437 } 2438 mergeFrom(com.google.protobuf.Message other)2439 public Builder mergeFrom(com.google.protobuf.Message other) { 2440 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse) { 2441 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse)other); 2442 } else { 2443 super.mergeFrom(other); 2444 return this; 2445 } 2446 } 2447 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse other)2448 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse other) { 2449 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance()) return this; 2450 this.mergeUnknownFields(other.getUnknownFields()); 2451 return this; 2452 } 2453 isInitialized()2454 public final boolean isInitialized() { 2455 return true; 2456 } 2457 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2458 public Builder mergeFrom( 2459 com.google.protobuf.CodedInputStream input, 2460 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2461 throws java.io.IOException { 2462 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse parsedMessage = null; 2463 try { 2464 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2465 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2466 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DeleteColumnResponse) e.getUnfinishedMessage(); 2467 throw e; 2468 } finally { 2469 if (parsedMessage != null) { 2470 mergeFrom(parsedMessage); 2471 } 2472 } 2473 return this; 2474 } 2475 2476 // @@protoc_insertion_point(builder_scope:DeleteColumnResponse) 2477 } 2478 2479 static { 2480 defaultInstance = new DeleteColumnResponse(true); defaultInstance.initFields()2481 defaultInstance.initFields(); 2482 } 2483 2484 // @@protoc_insertion_point(class_scope:DeleteColumnResponse) 2485 } 2486 2487 public interface ModifyColumnRequestOrBuilder 2488 extends com.google.protobuf.MessageOrBuilder { 2489 2490 // required .TableName table_name = 1; 2491 /** 2492 * <code>required .TableName table_name = 1;</code> 2493 */ hasTableName()2494 boolean hasTableName(); 2495 /** 2496 * <code>required .TableName table_name = 1;</code> 2497 */ getTableName()2498 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); 2499 /** 2500 * <code>required .TableName table_name = 1;</code> 2501 */ getTableNameOrBuilder()2502 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); 2503 2504 // required .ColumnFamilySchema column_families = 2; 2505 /** 2506 * <code>required .ColumnFamilySchema column_families = 2;</code> 2507 */ hasColumnFamilies()2508 boolean hasColumnFamilies(); 2509 /** 2510 * <code>required .ColumnFamilySchema column_families = 2;</code> 2511 */ getColumnFamilies()2512 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies(); 2513 /** 2514 * <code>required .ColumnFamilySchema column_families = 2;</code> 2515 */ getColumnFamiliesOrBuilder()2516 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder(); 2517 2518 // optional uint64 nonce_group = 3 [default = 0]; 2519 /** 2520 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2521 */ hasNonceGroup()2522 boolean hasNonceGroup(); 2523 /** 2524 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2525 */ getNonceGroup()2526 long getNonceGroup(); 2527 2528 // optional uint64 nonce = 4 [default = 0]; 2529 /** 2530 * <code>optional uint64 nonce = 4 [default = 0];</code> 2531 */ hasNonce()2532 boolean hasNonce(); 2533 /** 2534 * <code>optional uint64 nonce = 4 [default = 0];</code> 2535 */ getNonce()2536 long getNonce(); 2537 } 2538 /** 2539 * Protobuf type {@code ModifyColumnRequest} 2540 */ 2541 public static final class ModifyColumnRequest extends 2542 com.google.protobuf.GeneratedMessage 2543 implements ModifyColumnRequestOrBuilder { 2544 // Use ModifyColumnRequest.newBuilder() to construct. ModifyColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)2545 private ModifyColumnRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2546 super(builder); 2547 this.unknownFields = builder.getUnknownFields(); 2548 } ModifyColumnRequest(boolean noInit)2549 private ModifyColumnRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2550 2551 private static final ModifyColumnRequest defaultInstance; getDefaultInstance()2552 public static ModifyColumnRequest getDefaultInstance() { 2553 return defaultInstance; 2554 } 2555 getDefaultInstanceForType()2556 public ModifyColumnRequest getDefaultInstanceForType() { 2557 return defaultInstance; 2558 } 2559 2560 private final com.google.protobuf.UnknownFieldSet unknownFields; 2561 @java.lang.Override 2562 public final com.google.protobuf.UnknownFieldSet getUnknownFields()2563 getUnknownFields() { 2564 return this.unknownFields; 2565 } ModifyColumnRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2566 private ModifyColumnRequest( 2567 com.google.protobuf.CodedInputStream input, 2568 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2569 throws com.google.protobuf.InvalidProtocolBufferException { 2570 initFields(); 2571 int mutable_bitField0_ = 0; 2572 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2573 com.google.protobuf.UnknownFieldSet.newBuilder(); 2574 try { 2575 boolean done = false; 2576 while (!done) { 2577 int tag = input.readTag(); 2578 switch (tag) { 2579 case 0: 2580 done = true; 2581 break; 2582 default: { 2583 if (!parseUnknownField(input, unknownFields, 2584 extensionRegistry, tag)) { 2585 done = true; 2586 } 2587 break; 2588 } 2589 case 10: { 2590 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; 2591 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2592 subBuilder = tableName_.toBuilder(); 2593 } 2594 tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); 2595 if (subBuilder != null) { 2596 subBuilder.mergeFrom(tableName_); 2597 tableName_ = subBuilder.buildPartial(); 2598 } 2599 bitField0_ |= 0x00000001; 2600 break; 2601 } 2602 case 18: { 2603 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = null; 2604 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2605 subBuilder = columnFamilies_.toBuilder(); 2606 } 2607 columnFamilies_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry); 2608 if (subBuilder != null) { 2609 subBuilder.mergeFrom(columnFamilies_); 2610 columnFamilies_ = subBuilder.buildPartial(); 2611 } 2612 bitField0_ |= 0x00000002; 2613 break; 2614 } 2615 case 24: { 2616 bitField0_ |= 0x00000004; 2617 nonceGroup_ = input.readUInt64(); 2618 break; 2619 } 2620 case 32: { 2621 bitField0_ |= 0x00000008; 2622 nonce_ = input.readUInt64(); 2623 break; 2624 } 2625 } 2626 } 2627 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2628 throw e.setUnfinishedMessage(this); 2629 } catch (java.io.IOException e) { 2630 throw new com.google.protobuf.InvalidProtocolBufferException( 2631 e.getMessage()).setUnfinishedMessage(this); 2632 } finally { 2633 this.unknownFields = unknownFields.build(); 2634 makeExtensionsImmutable(); 2635 } 2636 } 2637 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()2638 getDescriptor() { 2639 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnRequest_descriptor; 2640 } 2641 2642 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()2643 internalGetFieldAccessorTable() { 2644 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnRequest_fieldAccessorTable 2645 .ensureFieldAccessorsInitialized( 2646 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.Builder.class); 2647 } 2648 2649 public static com.google.protobuf.Parser<ModifyColumnRequest> PARSER = 2650 new com.google.protobuf.AbstractParser<ModifyColumnRequest>() { 2651 public ModifyColumnRequest parsePartialFrom( 2652 com.google.protobuf.CodedInputStream input, 2653 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2654 throws com.google.protobuf.InvalidProtocolBufferException { 2655 return new ModifyColumnRequest(input, extensionRegistry); 2656 } 2657 }; 2658 2659 @java.lang.Override getParserForType()2660 public com.google.protobuf.Parser<ModifyColumnRequest> getParserForType() { 2661 return PARSER; 2662 } 2663 2664 private int bitField0_; 2665 // required .TableName table_name = 1; 2666 public static final int TABLE_NAME_FIELD_NUMBER = 1; 2667 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; 2668 /** 2669 * <code>required .TableName table_name = 1;</code> 2670 */ hasTableName()2671 public boolean hasTableName() { 2672 return ((bitField0_ & 0x00000001) == 0x00000001); 2673 } 2674 /** 2675 * <code>required .TableName table_name = 1;</code> 2676 */ getTableName()2677 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 2678 return tableName_; 2679 } 2680 /** 2681 * <code>required .TableName table_name = 1;</code> 2682 */ getTableNameOrBuilder()2683 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 2684 return tableName_; 2685 } 2686 2687 // required .ColumnFamilySchema column_families = 2; 2688 public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2; 2689 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_; 2690 /** 2691 * <code>required .ColumnFamilySchema column_families = 2;</code> 2692 */ hasColumnFamilies()2693 public boolean hasColumnFamilies() { 2694 return ((bitField0_ & 0x00000002) == 0x00000002); 2695 } 2696 /** 2697 * <code>required .ColumnFamilySchema column_families = 2;</code> 2698 */ getColumnFamilies()2699 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() { 2700 return columnFamilies_; 2701 } 2702 /** 2703 * <code>required .ColumnFamilySchema column_families = 2;</code> 2704 */ getColumnFamiliesOrBuilder()2705 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() { 2706 return columnFamilies_; 2707 } 2708 2709 // optional uint64 nonce_group = 3 [default = 0]; 2710 public static final int NONCE_GROUP_FIELD_NUMBER = 3; 2711 private long nonceGroup_; 2712 /** 2713 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2714 */ hasNonceGroup()2715 public boolean hasNonceGroup() { 2716 return ((bitField0_ & 0x00000004) == 0x00000004); 2717 } 2718 /** 2719 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 2720 */ getNonceGroup()2721 public long getNonceGroup() { 2722 return nonceGroup_; 2723 } 2724 2725 // optional uint64 nonce = 4 [default = 0]; 2726 public static final int NONCE_FIELD_NUMBER = 4; 2727 private long nonce_; 2728 /** 2729 * <code>optional uint64 nonce = 4 [default = 0];</code> 2730 */ hasNonce()2731 public boolean hasNonce() { 2732 return ((bitField0_ & 0x00000008) == 0x00000008); 2733 } 2734 /** 2735 * <code>optional uint64 nonce = 4 [default = 0];</code> 2736 */ getNonce()2737 public long getNonce() { 2738 return nonce_; 2739 } 2740 initFields()2741 private void initFields() { 2742 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 2743 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 2744 nonceGroup_ = 0L; 2745 nonce_ = 0L; 2746 } 2747 private byte memoizedIsInitialized = -1; isInitialized()2748 public final boolean isInitialized() { 2749 byte isInitialized = memoizedIsInitialized; 2750 if (isInitialized != -1) return isInitialized == 1; 2751 2752 if (!hasTableName()) { 2753 memoizedIsInitialized = 0; 2754 return false; 2755 } 2756 if (!hasColumnFamilies()) { 2757 memoizedIsInitialized = 0; 2758 return false; 2759 } 2760 if (!getTableName().isInitialized()) { 2761 memoizedIsInitialized = 0; 2762 return false; 2763 } 2764 if (!getColumnFamilies().isInitialized()) { 2765 memoizedIsInitialized = 0; 2766 return false; 2767 } 2768 memoizedIsInitialized = 1; 2769 return true; 2770 } 2771 writeTo(com.google.protobuf.CodedOutputStream output)2772 public void writeTo(com.google.protobuf.CodedOutputStream output) 2773 throws java.io.IOException { 2774 getSerializedSize(); 2775 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2776 output.writeMessage(1, tableName_); 2777 } 2778 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2779 output.writeMessage(2, columnFamilies_); 2780 } 2781 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2782 output.writeUInt64(3, nonceGroup_); 2783 } 2784 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2785 output.writeUInt64(4, nonce_); 2786 } 2787 getUnknownFields().writeTo(output); 2788 } 2789 2790 private int memoizedSerializedSize = -1; getSerializedSize()2791 public int getSerializedSize() { 2792 int size = memoizedSerializedSize; 2793 if (size != -1) return size; 2794 2795 size = 0; 2796 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2797 size += com.google.protobuf.CodedOutputStream 2798 .computeMessageSize(1, tableName_); 2799 } 2800 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2801 size += com.google.protobuf.CodedOutputStream 2802 .computeMessageSize(2, columnFamilies_); 2803 } 2804 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2805 size += com.google.protobuf.CodedOutputStream 2806 .computeUInt64Size(3, nonceGroup_); 2807 } 2808 if (((bitField0_ & 0x00000008) == 0x00000008)) { 2809 size += com.google.protobuf.CodedOutputStream 2810 .computeUInt64Size(4, nonce_); 2811 } 2812 size += getUnknownFields().getSerializedSize(); 2813 memoizedSerializedSize = size; 2814 return size; 2815 } 2816 2817 private static final long serialVersionUID = 0L; 2818 @java.lang.Override writeReplace()2819 protected java.lang.Object writeReplace() 2820 throws java.io.ObjectStreamException { 2821 return super.writeReplace(); 2822 } 2823 2824 @java.lang.Override equals(final java.lang.Object obj)2825 public boolean equals(final java.lang.Object obj) { 2826 if (obj == this) { 2827 return true; 2828 } 2829 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest)) { 2830 return super.equals(obj); 2831 } 2832 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest) obj; 2833 2834 boolean result = true; 2835 result = result && (hasTableName() == other.hasTableName()); 2836 if (hasTableName()) { 2837 result = result && getTableName() 2838 .equals(other.getTableName()); 2839 } 2840 result = result && (hasColumnFamilies() == other.hasColumnFamilies()); 2841 if (hasColumnFamilies()) { 2842 result = result && getColumnFamilies() 2843 .equals(other.getColumnFamilies()); 2844 } 2845 result = result && (hasNonceGroup() == other.hasNonceGroup()); 2846 if (hasNonceGroup()) { 2847 result = result && (getNonceGroup() 2848 == other.getNonceGroup()); 2849 } 2850 result = result && (hasNonce() == other.hasNonce()); 2851 if (hasNonce()) { 2852 result = result && (getNonce() 2853 == other.getNonce()); 2854 } 2855 result = result && 2856 getUnknownFields().equals(other.getUnknownFields()); 2857 return result; 2858 } 2859 2860 private int memoizedHashCode = 0; 2861 @java.lang.Override hashCode()2862 public int hashCode() { 2863 if (memoizedHashCode != 0) { 2864 return memoizedHashCode; 2865 } 2866 int hash = 41; 2867 hash = (19 * hash) + getDescriptorForType().hashCode(); 2868 if (hasTableName()) { 2869 hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; 2870 hash = (53 * hash) + getTableName().hashCode(); 2871 } 2872 if (hasColumnFamilies()) { 2873 hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER; 2874 hash = (53 * hash) + getColumnFamilies().hashCode(); 2875 } 2876 if (hasNonceGroup()) { 2877 hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; 2878 hash = (53 * hash) + hashLong(getNonceGroup()); 2879 } 2880 if (hasNonce()) { 2881 hash = (37 * hash) + NONCE_FIELD_NUMBER; 2882 hash = (53 * hash) + hashLong(getNonce()); 2883 } 2884 hash = (29 * hash) + getUnknownFields().hashCode(); 2885 memoizedHashCode = hash; 2886 return hash; 2887 } 2888 parseFrom( com.google.protobuf.ByteString data)2889 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2890 com.google.protobuf.ByteString data) 2891 throws com.google.protobuf.InvalidProtocolBufferException { 2892 return PARSER.parseFrom(data); 2893 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2894 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2895 com.google.protobuf.ByteString data, 2896 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2897 throws com.google.protobuf.InvalidProtocolBufferException { 2898 return PARSER.parseFrom(data, extensionRegistry); 2899 } parseFrom(byte[] data)2900 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(byte[] data) 2901 throws com.google.protobuf.InvalidProtocolBufferException { 2902 return PARSER.parseFrom(data); 2903 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2904 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2905 byte[] data, 2906 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2907 throws com.google.protobuf.InvalidProtocolBufferException { 2908 return PARSER.parseFrom(data, extensionRegistry); 2909 } parseFrom(java.io.InputStream input)2910 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(java.io.InputStream input) 2911 throws java.io.IOException { 2912 return PARSER.parseFrom(input); 2913 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2914 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2915 java.io.InputStream input, 2916 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2917 throws java.io.IOException { 2918 return PARSER.parseFrom(input, extensionRegistry); 2919 } parseDelimitedFrom(java.io.InputStream input)2920 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseDelimitedFrom(java.io.InputStream input) 2921 throws java.io.IOException { 2922 return PARSER.parseDelimitedFrom(input); 2923 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2924 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseDelimitedFrom( 2925 java.io.InputStream input, 2926 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2927 throws java.io.IOException { 2928 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2929 } parseFrom( com.google.protobuf.CodedInputStream input)2930 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2931 com.google.protobuf.CodedInputStream input) 2932 throws java.io.IOException { 2933 return PARSER.parseFrom(input); 2934 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)2935 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom( 2936 com.google.protobuf.CodedInputStream input, 2937 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2938 throws java.io.IOException { 2939 return PARSER.parseFrom(input, extensionRegistry); 2940 } 2941 newBuilder()2942 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()2943 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest prototype)2944 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest prototype) { 2945 return newBuilder().mergeFrom(prototype); 2946 } toBuilder()2947 public Builder toBuilder() { return newBuilder(this); } 2948 2949 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)2950 protected Builder newBuilderForType( 2951 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2952 Builder builder = new Builder(parent); 2953 return builder; 2954 } 2955 /** 2956 * Protobuf type {@code ModifyColumnRequest} 2957 */ 2958 public static final class Builder extends 2959 com.google.protobuf.GeneratedMessage.Builder<Builder> 2960 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequestOrBuilder { 2961 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()2962 getDescriptor() { 2963 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnRequest_descriptor; 2964 } 2965 2966 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()2967 internalGetFieldAccessorTable() { 2968 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnRequest_fieldAccessorTable 2969 .ensureFieldAccessorsInitialized( 2970 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.Builder.class); 2971 } 2972 2973 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.newBuilder() Builder()2974 private Builder() { 2975 maybeForceBuilderInitialization(); 2976 } 2977 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)2978 private Builder( 2979 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2980 super(parent); 2981 maybeForceBuilderInitialization(); 2982 } maybeForceBuilderInitialization()2983 private void maybeForceBuilderInitialization() { 2984 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2985 getTableNameFieldBuilder(); 2986 getColumnFamiliesFieldBuilder(); 2987 } 2988 } create()2989 private static Builder create() { 2990 return new Builder(); 2991 } 2992 clear()2993 public Builder clear() { 2994 super.clear(); 2995 if (tableNameBuilder_ == null) { 2996 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 2997 } else { 2998 tableNameBuilder_.clear(); 2999 } 3000 bitField0_ = (bitField0_ & ~0x00000001); 3001 if (columnFamiliesBuilder_ == null) { 3002 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 3003 } else { 3004 columnFamiliesBuilder_.clear(); 3005 } 3006 bitField0_ = (bitField0_ & ~0x00000002); 3007 nonceGroup_ = 0L; 3008 bitField0_ = (bitField0_ & ~0x00000004); 3009 nonce_ = 0L; 3010 bitField0_ = (bitField0_ & ~0x00000008); 3011 return this; 3012 } 3013 clone()3014 public Builder clone() { 3015 return create().mergeFrom(buildPartial()); 3016 } 3017 3018 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()3019 getDescriptorForType() { 3020 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnRequest_descriptor; 3021 } 3022 getDefaultInstanceForType()3023 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest getDefaultInstanceForType() { 3024 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance(); 3025 } 3026 build()3027 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest build() { 3028 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest result = buildPartial(); 3029 if (!result.isInitialized()) { 3030 throw newUninitializedMessageException(result); 3031 } 3032 return result; 3033 } 3034 buildPartial()3035 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest buildPartial() { 3036 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest(this); 3037 int from_bitField0_ = bitField0_; 3038 int to_bitField0_ = 0; 3039 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 3040 to_bitField0_ |= 0x00000001; 3041 } 3042 if (tableNameBuilder_ == null) { 3043 result.tableName_ = tableName_; 3044 } else { 3045 result.tableName_ = tableNameBuilder_.build(); 3046 } 3047 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 3048 to_bitField0_ |= 0x00000002; 3049 } 3050 if (columnFamiliesBuilder_ == null) { 3051 result.columnFamilies_ = columnFamilies_; 3052 } else { 3053 result.columnFamilies_ = columnFamiliesBuilder_.build(); 3054 } 3055 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 3056 to_bitField0_ |= 0x00000004; 3057 } 3058 result.nonceGroup_ = nonceGroup_; 3059 if (((from_bitField0_ & 0x00000008) == 0x00000008)) { 3060 to_bitField0_ |= 0x00000008; 3061 } 3062 result.nonce_ = nonce_; 3063 result.bitField0_ = to_bitField0_; 3064 onBuilt(); 3065 return result; 3066 } 3067 mergeFrom(com.google.protobuf.Message other)3068 public Builder mergeFrom(com.google.protobuf.Message other) { 3069 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest) { 3070 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest)other); 3071 } else { 3072 super.mergeFrom(other); 3073 return this; 3074 } 3075 } 3076 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest other)3077 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest other) { 3078 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance()) return this; 3079 if (other.hasTableName()) { 3080 mergeTableName(other.getTableName()); 3081 } 3082 if (other.hasColumnFamilies()) { 3083 mergeColumnFamilies(other.getColumnFamilies()); 3084 } 3085 if (other.hasNonceGroup()) { 3086 setNonceGroup(other.getNonceGroup()); 3087 } 3088 if (other.hasNonce()) { 3089 setNonce(other.getNonce()); 3090 } 3091 this.mergeUnknownFields(other.getUnknownFields()); 3092 return this; 3093 } 3094 isInitialized()3095 public final boolean isInitialized() { 3096 if (!hasTableName()) { 3097 3098 return false; 3099 } 3100 if (!hasColumnFamilies()) { 3101 3102 return false; 3103 } 3104 if (!getTableName().isInitialized()) { 3105 3106 return false; 3107 } 3108 if (!getColumnFamilies().isInitialized()) { 3109 3110 return false; 3111 } 3112 return true; 3113 } 3114 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3115 public Builder mergeFrom( 3116 com.google.protobuf.CodedInputStream input, 3117 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3118 throws java.io.IOException { 3119 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest parsedMessage = null; 3120 try { 3121 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3122 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3123 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnRequest) e.getUnfinishedMessage(); 3124 throw e; 3125 } finally { 3126 if (parsedMessage != null) { 3127 mergeFrom(parsedMessage); 3128 } 3129 } 3130 return this; 3131 } 3132 private int bitField0_; 3133 3134 // required .TableName table_name = 1; 3135 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 3136 private com.google.protobuf.SingleFieldBuilder< 3137 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; 3138 /** 3139 * <code>required .TableName table_name = 1;</code> 3140 */ hasTableName()3141 public boolean hasTableName() { 3142 return ((bitField0_ & 0x00000001) == 0x00000001); 3143 } 3144 /** 3145 * <code>required .TableName table_name = 1;</code> 3146 */ getTableName()3147 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { 3148 if (tableNameBuilder_ == null) { 3149 return tableName_; 3150 } else { 3151 return tableNameBuilder_.getMessage(); 3152 } 3153 } 3154 /** 3155 * <code>required .TableName table_name = 1;</code> 3156 */ setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)3157 public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 3158 if (tableNameBuilder_ == null) { 3159 if (value == null) { 3160 throw new NullPointerException(); 3161 } 3162 tableName_ = value; 3163 onChanged(); 3164 } else { 3165 tableNameBuilder_.setMessage(value); 3166 } 3167 bitField0_ |= 0x00000001; 3168 return this; 3169 } 3170 /** 3171 * <code>required .TableName table_name = 1;</code> 3172 */ setTableName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue)3173 public Builder setTableName( 3174 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { 3175 if (tableNameBuilder_ == null) { 3176 tableName_ = builderForValue.build(); 3177 onChanged(); 3178 } else { 3179 tableNameBuilder_.setMessage(builderForValue.build()); 3180 } 3181 bitField0_ |= 0x00000001; 3182 return this; 3183 } 3184 /** 3185 * <code>required .TableName table_name = 1;</code> 3186 */ mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value)3187 public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { 3188 if (tableNameBuilder_ == null) { 3189 if (((bitField0_ & 0x00000001) == 0x00000001) && 3190 tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { 3191 tableName_ = 3192 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); 3193 } else { 3194 tableName_ = value; 3195 } 3196 onChanged(); 3197 } else { 3198 tableNameBuilder_.mergeFrom(value); 3199 } 3200 bitField0_ |= 0x00000001; 3201 return this; 3202 } 3203 /** 3204 * <code>required .TableName table_name = 1;</code> 3205 */ clearTableName()3206 public Builder clearTableName() { 3207 if (tableNameBuilder_ == null) { 3208 tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); 3209 onChanged(); 3210 } else { 3211 tableNameBuilder_.clear(); 3212 } 3213 bitField0_ = (bitField0_ & ~0x00000001); 3214 return this; 3215 } 3216 /** 3217 * <code>required .TableName table_name = 1;</code> 3218 */ getTableNameBuilder()3219 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { 3220 bitField0_ |= 0x00000001; 3221 onChanged(); 3222 return getTableNameFieldBuilder().getBuilder(); 3223 } 3224 /** 3225 * <code>required .TableName table_name = 1;</code> 3226 */ getTableNameOrBuilder()3227 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { 3228 if (tableNameBuilder_ != null) { 3229 return tableNameBuilder_.getMessageOrBuilder(); 3230 } else { 3231 return tableName_; 3232 } 3233 } 3234 /** 3235 * <code>required .TableName table_name = 1;</code> 3236 */ 3237 private com.google.protobuf.SingleFieldBuilder< 3238 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder()3239 getTableNameFieldBuilder() { 3240 if (tableNameBuilder_ == null) { 3241 tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< 3242 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( 3243 tableName_, 3244 getParentForChildren(), 3245 isClean()); 3246 tableName_ = null; 3247 } 3248 return tableNameBuilder_; 3249 } 3250 3251 // required .ColumnFamilySchema column_families = 2; 3252 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 3253 private com.google.protobuf.SingleFieldBuilder< 3254 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_; 3255 /** 3256 * <code>required .ColumnFamilySchema column_families = 2;</code> 3257 */ hasColumnFamilies()3258 public boolean hasColumnFamilies() { 3259 return ((bitField0_ & 0x00000002) == 0x00000002); 3260 } 3261 /** 3262 * <code>required .ColumnFamilySchema column_families = 2;</code> 3263 */ getColumnFamilies()3264 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() { 3265 if (columnFamiliesBuilder_ == null) { 3266 return columnFamilies_; 3267 } else { 3268 return columnFamiliesBuilder_.getMessage(); 3269 } 3270 } 3271 /** 3272 * <code>required .ColumnFamilySchema column_families = 2;</code> 3273 */ setColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value)3274 public Builder setColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { 3275 if (columnFamiliesBuilder_ == null) { 3276 if (value == null) { 3277 throw new NullPointerException(); 3278 } 3279 columnFamilies_ = value; 3280 onChanged(); 3281 } else { 3282 columnFamiliesBuilder_.setMessage(value); 3283 } 3284 bitField0_ |= 0x00000002; 3285 return this; 3286 } 3287 /** 3288 * <code>required .ColumnFamilySchema column_families = 2;</code> 3289 */ setColumnFamilies( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue)3290 public Builder setColumnFamilies( 3291 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) { 3292 if (columnFamiliesBuilder_ == null) { 3293 columnFamilies_ = builderForValue.build(); 3294 onChanged(); 3295 } else { 3296 columnFamiliesBuilder_.setMessage(builderForValue.build()); 3297 } 3298 bitField0_ |= 0x00000002; 3299 return this; 3300 } 3301 /** 3302 * <code>required .ColumnFamilySchema column_families = 2;</code> 3303 */ mergeColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value)3304 public Builder mergeColumnFamilies(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema value) { 3305 if (columnFamiliesBuilder_ == null) { 3306 if (((bitField0_ & 0x00000002) == 0x00000002) && 3307 columnFamilies_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) { 3308 columnFamilies_ = 3309 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(columnFamilies_).mergeFrom(value).buildPartial(); 3310 } else { 3311 columnFamilies_ = value; 3312 } 3313 onChanged(); 3314 } else { 3315 columnFamiliesBuilder_.mergeFrom(value); 3316 } 3317 bitField0_ |= 0x00000002; 3318 return this; 3319 } 3320 /** 3321 * <code>required .ColumnFamilySchema column_families = 2;</code> 3322 */ clearColumnFamilies()3323 public Builder clearColumnFamilies() { 3324 if (columnFamiliesBuilder_ == null) { 3325 columnFamilies_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance(); 3326 onChanged(); 3327 } else { 3328 columnFamiliesBuilder_.clear(); 3329 } 3330 bitField0_ = (bitField0_ & ~0x00000002); 3331 return this; 3332 } 3333 /** 3334 * <code>required .ColumnFamilySchema column_families = 2;</code> 3335 */ getColumnFamiliesBuilder()3336 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder() { 3337 bitField0_ |= 0x00000002; 3338 onChanged(); 3339 return getColumnFamiliesFieldBuilder().getBuilder(); 3340 } 3341 /** 3342 * <code>required .ColumnFamilySchema column_families = 2;</code> 3343 */ getColumnFamiliesOrBuilder()3344 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() { 3345 if (columnFamiliesBuilder_ != null) { 3346 return columnFamiliesBuilder_.getMessageOrBuilder(); 3347 } else { 3348 return columnFamilies_; 3349 } 3350 } 3351 /** 3352 * <code>required .ColumnFamilySchema column_families = 2;</code> 3353 */ 3354 private com.google.protobuf.SingleFieldBuilder< 3355 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> getColumnFamiliesFieldBuilder()3356 getColumnFamiliesFieldBuilder() { 3357 if (columnFamiliesBuilder_ == null) { 3358 columnFamiliesBuilder_ = new com.google.protobuf.SingleFieldBuilder< 3359 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>( 3360 columnFamilies_, 3361 getParentForChildren(), 3362 isClean()); 3363 columnFamilies_ = null; 3364 } 3365 return columnFamiliesBuilder_; 3366 } 3367 3368 // optional uint64 nonce_group = 3 [default = 0]; 3369 private long nonceGroup_ ; 3370 /** 3371 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 3372 */ hasNonceGroup()3373 public boolean hasNonceGroup() { 3374 return ((bitField0_ & 0x00000004) == 0x00000004); 3375 } 3376 /** 3377 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 3378 */ getNonceGroup()3379 public long getNonceGroup() { 3380 return nonceGroup_; 3381 } 3382 /** 3383 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 3384 */ setNonceGroup(long value)3385 public Builder setNonceGroup(long value) { 3386 bitField0_ |= 0x00000004; 3387 nonceGroup_ = value; 3388 onChanged(); 3389 return this; 3390 } 3391 /** 3392 * <code>optional uint64 nonce_group = 3 [default = 0];</code> 3393 */ clearNonceGroup()3394 public Builder clearNonceGroup() { 3395 bitField0_ = (bitField0_ & ~0x00000004); 3396 nonceGroup_ = 0L; 3397 onChanged(); 3398 return this; 3399 } 3400 3401 // optional uint64 nonce = 4 [default = 0]; 3402 private long nonce_ ; 3403 /** 3404 * <code>optional uint64 nonce = 4 [default = 0];</code> 3405 */ hasNonce()3406 public boolean hasNonce() { 3407 return ((bitField0_ & 0x00000008) == 0x00000008); 3408 } 3409 /** 3410 * <code>optional uint64 nonce = 4 [default = 0];</code> 3411 */ getNonce()3412 public long getNonce() { 3413 return nonce_; 3414 } 3415 /** 3416 * <code>optional uint64 nonce = 4 [default = 0];</code> 3417 */ setNonce(long value)3418 public Builder setNonce(long value) { 3419 bitField0_ |= 0x00000008; 3420 nonce_ = value; 3421 onChanged(); 3422 return this; 3423 } 3424 /** 3425 * <code>optional uint64 nonce = 4 [default = 0];</code> 3426 */ clearNonce()3427 public Builder clearNonce() { 3428 bitField0_ = (bitField0_ & ~0x00000008); 3429 nonce_ = 0L; 3430 onChanged(); 3431 return this; 3432 } 3433 3434 // @@protoc_insertion_point(builder_scope:ModifyColumnRequest) 3435 } 3436 3437 static { 3438 defaultInstance = new ModifyColumnRequest(true); defaultInstance.initFields()3439 defaultInstance.initFields(); 3440 } 3441 3442 // @@protoc_insertion_point(class_scope:ModifyColumnRequest) 3443 } 3444 3445 public interface ModifyColumnResponseOrBuilder 3446 extends com.google.protobuf.MessageOrBuilder { 3447 } 3448 /** 3449 * Protobuf type {@code ModifyColumnResponse} 3450 */ 3451 public static final class ModifyColumnResponse extends 3452 com.google.protobuf.GeneratedMessage 3453 implements ModifyColumnResponseOrBuilder { 3454 // Use ModifyColumnResponse.newBuilder() to construct. ModifyColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)3455 private ModifyColumnResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3456 super(builder); 3457 this.unknownFields = builder.getUnknownFields(); 3458 } ModifyColumnResponse(boolean noInit)3459 private ModifyColumnResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3460 3461 private static final ModifyColumnResponse defaultInstance; getDefaultInstance()3462 public static ModifyColumnResponse getDefaultInstance() { 3463 return defaultInstance; 3464 } 3465 getDefaultInstanceForType()3466 public ModifyColumnResponse getDefaultInstanceForType() { 3467 return defaultInstance; 3468 } 3469 3470 private final com.google.protobuf.UnknownFieldSet unknownFields; 3471 @java.lang.Override 3472 public final com.google.protobuf.UnknownFieldSet getUnknownFields()3473 getUnknownFields() { 3474 return this.unknownFields; 3475 } ModifyColumnResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3476 private ModifyColumnResponse( 3477 com.google.protobuf.CodedInputStream input, 3478 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3479 throws com.google.protobuf.InvalidProtocolBufferException { 3480 initFields(); 3481 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3482 com.google.protobuf.UnknownFieldSet.newBuilder(); 3483 try { 3484 boolean done = false; 3485 while (!done) { 3486 int tag = input.readTag(); 3487 switch (tag) { 3488 case 0: 3489 done = true; 3490 break; 3491 default: { 3492 if (!parseUnknownField(input, unknownFields, 3493 extensionRegistry, tag)) { 3494 done = true; 3495 } 3496 break; 3497 } 3498 } 3499 } 3500 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3501 throw e.setUnfinishedMessage(this); 3502 } catch (java.io.IOException e) { 3503 throw new com.google.protobuf.InvalidProtocolBufferException( 3504 e.getMessage()).setUnfinishedMessage(this); 3505 } finally { 3506 this.unknownFields = unknownFields.build(); 3507 makeExtensionsImmutable(); 3508 } 3509 } 3510 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()3511 getDescriptor() { 3512 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnResponse_descriptor; 3513 } 3514 3515 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()3516 internalGetFieldAccessorTable() { 3517 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnResponse_fieldAccessorTable 3518 .ensureFieldAccessorsInitialized( 3519 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.Builder.class); 3520 } 3521 3522 public static com.google.protobuf.Parser<ModifyColumnResponse> PARSER = 3523 new com.google.protobuf.AbstractParser<ModifyColumnResponse>() { 3524 public ModifyColumnResponse parsePartialFrom( 3525 com.google.protobuf.CodedInputStream input, 3526 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3527 throws com.google.protobuf.InvalidProtocolBufferException { 3528 return new ModifyColumnResponse(input, extensionRegistry); 3529 } 3530 }; 3531 3532 @java.lang.Override getParserForType()3533 public com.google.protobuf.Parser<ModifyColumnResponse> getParserForType() { 3534 return PARSER; 3535 } 3536 initFields()3537 private void initFields() { 3538 } 3539 private byte memoizedIsInitialized = -1; isInitialized()3540 public final boolean isInitialized() { 3541 byte isInitialized = memoizedIsInitialized; 3542 if (isInitialized != -1) return isInitialized == 1; 3543 3544 memoizedIsInitialized = 1; 3545 return true; 3546 } 3547 writeTo(com.google.protobuf.CodedOutputStream output)3548 public void writeTo(com.google.protobuf.CodedOutputStream output) 3549 throws java.io.IOException { 3550 getSerializedSize(); 3551 getUnknownFields().writeTo(output); 3552 } 3553 3554 private int memoizedSerializedSize = -1; getSerializedSize()3555 public int getSerializedSize() { 3556 int size = memoizedSerializedSize; 3557 if (size != -1) return size; 3558 3559 size = 0; 3560 size += getUnknownFields().getSerializedSize(); 3561 memoizedSerializedSize = size; 3562 return size; 3563 } 3564 3565 private static final long serialVersionUID = 0L; 3566 @java.lang.Override writeReplace()3567 protected java.lang.Object writeReplace() 3568 throws java.io.ObjectStreamException { 3569 return super.writeReplace(); 3570 } 3571 3572 @java.lang.Override equals(final java.lang.Object obj)3573 public boolean equals(final java.lang.Object obj) { 3574 if (obj == this) { 3575 return true; 3576 } 3577 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse)) { 3578 return super.equals(obj); 3579 } 3580 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse) obj; 3581 3582 boolean result = true; 3583 result = result && 3584 getUnknownFields().equals(other.getUnknownFields()); 3585 return result; 3586 } 3587 3588 private int memoizedHashCode = 0; 3589 @java.lang.Override hashCode()3590 public int hashCode() { 3591 if (memoizedHashCode != 0) { 3592 return memoizedHashCode; 3593 } 3594 int hash = 41; 3595 hash = (19 * hash) + getDescriptorForType().hashCode(); 3596 hash = (29 * hash) + getUnknownFields().hashCode(); 3597 memoizedHashCode = hash; 3598 return hash; 3599 } 3600 parseFrom( com.google.protobuf.ByteString data)3601 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3602 com.google.protobuf.ByteString data) 3603 throws com.google.protobuf.InvalidProtocolBufferException { 3604 return PARSER.parseFrom(data); 3605 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3606 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3607 com.google.protobuf.ByteString data, 3608 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3609 throws com.google.protobuf.InvalidProtocolBufferException { 3610 return PARSER.parseFrom(data, extensionRegistry); 3611 } parseFrom(byte[] data)3612 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(byte[] data) 3613 throws com.google.protobuf.InvalidProtocolBufferException { 3614 return PARSER.parseFrom(data); 3615 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3616 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3617 byte[] data, 3618 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3619 throws com.google.protobuf.InvalidProtocolBufferException { 3620 return PARSER.parseFrom(data, extensionRegistry); 3621 } parseFrom(java.io.InputStream input)3622 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(java.io.InputStream input) 3623 throws java.io.IOException { 3624 return PARSER.parseFrom(input); 3625 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3626 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3627 java.io.InputStream input, 3628 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3629 throws java.io.IOException { 3630 return PARSER.parseFrom(input, extensionRegistry); 3631 } parseDelimitedFrom(java.io.InputStream input)3632 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseDelimitedFrom(java.io.InputStream input) 3633 throws java.io.IOException { 3634 return PARSER.parseDelimitedFrom(input); 3635 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3636 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseDelimitedFrom( 3637 java.io.InputStream input, 3638 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3639 throws java.io.IOException { 3640 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3641 } parseFrom( com.google.protobuf.CodedInputStream input)3642 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3643 com.google.protobuf.CodedInputStream input) 3644 throws java.io.IOException { 3645 return PARSER.parseFrom(input); 3646 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3647 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom( 3648 com.google.protobuf.CodedInputStream input, 3649 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3650 throws java.io.IOException { 3651 return PARSER.parseFrom(input, extensionRegistry); 3652 } 3653 newBuilder()3654 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()3655 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse prototype)3656 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse prototype) { 3657 return newBuilder().mergeFrom(prototype); 3658 } toBuilder()3659 public Builder toBuilder() { return newBuilder(this); } 3660 3661 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)3662 protected Builder newBuilderForType( 3663 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3664 Builder builder = new Builder(parent); 3665 return builder; 3666 } 3667 /** 3668 * Protobuf type {@code ModifyColumnResponse} 3669 */ 3670 public static final class Builder extends 3671 com.google.protobuf.GeneratedMessage.Builder<Builder> 3672 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponseOrBuilder { 3673 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()3674 getDescriptor() { 3675 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnResponse_descriptor; 3676 } 3677 3678 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()3679 internalGetFieldAccessorTable() { 3680 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnResponse_fieldAccessorTable 3681 .ensureFieldAccessorsInitialized( 3682 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.Builder.class); 3683 } 3684 3685 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.newBuilder() Builder()3686 private Builder() { 3687 maybeForceBuilderInitialization(); 3688 } 3689 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)3690 private Builder( 3691 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3692 super(parent); 3693 maybeForceBuilderInitialization(); 3694 } maybeForceBuilderInitialization()3695 private void maybeForceBuilderInitialization() { 3696 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3697 } 3698 } create()3699 private static Builder create() { 3700 return new Builder(); 3701 } 3702 clear()3703 public Builder clear() { 3704 super.clear(); 3705 return this; 3706 } 3707 clone()3708 public Builder clone() { 3709 return create().mergeFrom(buildPartial()); 3710 } 3711 3712 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()3713 getDescriptorForType() { 3714 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_ModifyColumnResponse_descriptor; 3715 } 3716 getDefaultInstanceForType()3717 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse getDefaultInstanceForType() { 3718 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(); 3719 } 3720 build()3721 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse build() { 3722 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse result = buildPartial(); 3723 if (!result.isInitialized()) { 3724 throw newUninitializedMessageException(result); 3725 } 3726 return result; 3727 } 3728 buildPartial()3729 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse buildPartial() { 3730 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse(this); 3731 onBuilt(); 3732 return result; 3733 } 3734 mergeFrom(com.google.protobuf.Message other)3735 public Builder mergeFrom(com.google.protobuf.Message other) { 3736 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse) { 3737 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse)other); 3738 } else { 3739 super.mergeFrom(other); 3740 return this; 3741 } 3742 } 3743 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse other)3744 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse other) { 3745 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance()) return this; 3746 this.mergeUnknownFields(other.getUnknownFields()); 3747 return this; 3748 } 3749 isInitialized()3750 public final boolean isInitialized() { 3751 return true; 3752 } 3753 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3754 public Builder mergeFrom( 3755 com.google.protobuf.CodedInputStream input, 3756 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3757 throws java.io.IOException { 3758 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse parsedMessage = null; 3759 try { 3760 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3761 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3762 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.ModifyColumnResponse) e.getUnfinishedMessage(); 3763 throw e; 3764 } finally { 3765 if (parsedMessage != null) { 3766 mergeFrom(parsedMessage); 3767 } 3768 } 3769 return this; 3770 } 3771 3772 // @@protoc_insertion_point(builder_scope:ModifyColumnResponse) 3773 } 3774 3775 static { 3776 defaultInstance = new ModifyColumnResponse(true); defaultInstance.initFields()3777 defaultInstance.initFields(); 3778 } 3779 3780 // @@protoc_insertion_point(class_scope:ModifyColumnResponse) 3781 } 3782 3783 public interface MoveRegionRequestOrBuilder 3784 extends com.google.protobuf.MessageOrBuilder { 3785 3786 // required .RegionSpecifier region = 1; 3787 /** 3788 * <code>required .RegionSpecifier region = 1;</code> 3789 */ hasRegion()3790 boolean hasRegion(); 3791 /** 3792 * <code>required .RegionSpecifier region = 1;</code> 3793 */ getRegion()3794 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); 3795 /** 3796 * <code>required .RegionSpecifier region = 1;</code> 3797 */ getRegionOrBuilder()3798 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); 3799 3800 // optional .ServerName dest_server_name = 2; 3801 /** 3802 * <code>optional .ServerName dest_server_name = 2;</code> 3803 */ hasDestServerName()3804 boolean hasDestServerName(); 3805 /** 3806 * <code>optional .ServerName dest_server_name = 2;</code> 3807 */ getDestServerName()3808 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName(); 3809 /** 3810 * <code>optional .ServerName dest_server_name = 2;</code> 3811 */ getDestServerNameOrBuilder()3812 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder(); 3813 } 3814 /** 3815 * Protobuf type {@code MoveRegionRequest} 3816 */ 3817 public static final class MoveRegionRequest extends 3818 com.google.protobuf.GeneratedMessage 3819 implements MoveRegionRequestOrBuilder { 3820 // Use MoveRegionRequest.newBuilder() to construct. MoveRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)3821 private MoveRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3822 super(builder); 3823 this.unknownFields = builder.getUnknownFields(); 3824 } MoveRegionRequest(boolean noInit)3825 private MoveRegionRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3826 3827 private static final MoveRegionRequest defaultInstance; getDefaultInstance()3828 public static MoveRegionRequest getDefaultInstance() { 3829 return defaultInstance; 3830 } 3831 getDefaultInstanceForType()3832 public MoveRegionRequest getDefaultInstanceForType() { 3833 return defaultInstance; 3834 } 3835 3836 private final com.google.protobuf.UnknownFieldSet unknownFields; 3837 @java.lang.Override 3838 public final com.google.protobuf.UnknownFieldSet getUnknownFields()3839 getUnknownFields() { 3840 return this.unknownFields; 3841 } MoveRegionRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)3842 private MoveRegionRequest( 3843 com.google.protobuf.CodedInputStream input, 3844 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3845 throws com.google.protobuf.InvalidProtocolBufferException { 3846 initFields(); 3847 int mutable_bitField0_ = 0; 3848 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3849 com.google.protobuf.UnknownFieldSet.newBuilder(); 3850 try { 3851 boolean done = false; 3852 while (!done) { 3853 int tag = input.readTag(); 3854 switch (tag) { 3855 case 0: 3856 done = true; 3857 break; 3858 default: { 3859 if (!parseUnknownField(input, unknownFields, 3860 extensionRegistry, tag)) { 3861 done = true; 3862 } 3863 break; 3864 } 3865 case 10: { 3866 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 3867 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3868 subBuilder = region_.toBuilder(); 3869 } 3870 region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 3871 if (subBuilder != null) { 3872 subBuilder.mergeFrom(region_); 3873 region_ = subBuilder.buildPartial(); 3874 } 3875 bitField0_ |= 0x00000001; 3876 break; 3877 } 3878 case 18: { 3879 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; 3880 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3881 subBuilder = destServerName_.toBuilder(); 3882 } 3883 destServerName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); 3884 if (subBuilder != null) { 3885 subBuilder.mergeFrom(destServerName_); 3886 destServerName_ = subBuilder.buildPartial(); 3887 } 3888 bitField0_ |= 0x00000002; 3889 break; 3890 } 3891 } 3892 } 3893 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3894 throw e.setUnfinishedMessage(this); 3895 } catch (java.io.IOException e) { 3896 throw new com.google.protobuf.InvalidProtocolBufferException( 3897 e.getMessage()).setUnfinishedMessage(this); 3898 } finally { 3899 this.unknownFields = unknownFields.build(); 3900 makeExtensionsImmutable(); 3901 } 3902 } 3903 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()3904 getDescriptor() { 3905 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionRequest_descriptor; 3906 } 3907 3908 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()3909 internalGetFieldAccessorTable() { 3910 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionRequest_fieldAccessorTable 3911 .ensureFieldAccessorsInitialized( 3912 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.Builder.class); 3913 } 3914 3915 public static com.google.protobuf.Parser<MoveRegionRequest> PARSER = 3916 new com.google.protobuf.AbstractParser<MoveRegionRequest>() { 3917 public MoveRegionRequest parsePartialFrom( 3918 com.google.protobuf.CodedInputStream input, 3919 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3920 throws com.google.protobuf.InvalidProtocolBufferException { 3921 return new MoveRegionRequest(input, extensionRegistry); 3922 } 3923 }; 3924 3925 @java.lang.Override getParserForType()3926 public com.google.protobuf.Parser<MoveRegionRequest> getParserForType() { 3927 return PARSER; 3928 } 3929 3930 private int bitField0_; 3931 // required .RegionSpecifier region = 1; 3932 public static final int REGION_FIELD_NUMBER = 1; 3933 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_; 3934 /** 3935 * <code>required .RegionSpecifier region = 1;</code> 3936 */ hasRegion()3937 public boolean hasRegion() { 3938 return ((bitField0_ & 0x00000001) == 0x00000001); 3939 } 3940 /** 3941 * <code>required .RegionSpecifier region = 1;</code> 3942 */ getRegion()3943 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 3944 return region_; 3945 } 3946 /** 3947 * <code>required .RegionSpecifier region = 1;</code> 3948 */ getRegionOrBuilder()3949 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 3950 return region_; 3951 } 3952 3953 // optional .ServerName dest_server_name = 2; 3954 public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2; 3955 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_; 3956 /** 3957 * <code>optional .ServerName dest_server_name = 2;</code> 3958 */ hasDestServerName()3959 public boolean hasDestServerName() { 3960 return ((bitField0_ & 0x00000002) == 0x00000002); 3961 } 3962 /** 3963 * <code>optional .ServerName dest_server_name = 2;</code> 3964 */ getDestServerName()3965 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName() { 3966 return destServerName_; 3967 } 3968 /** 3969 * <code>optional .ServerName dest_server_name = 2;</code> 3970 */ getDestServerNameOrBuilder()3971 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder() { 3972 return destServerName_; 3973 } 3974 initFields()3975 private void initFields() { 3976 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 3977 destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); 3978 } 3979 private byte memoizedIsInitialized = -1; isInitialized()3980 public final boolean isInitialized() { 3981 byte isInitialized = memoizedIsInitialized; 3982 if (isInitialized != -1) return isInitialized == 1; 3983 3984 if (!hasRegion()) { 3985 memoizedIsInitialized = 0; 3986 return false; 3987 } 3988 if (!getRegion().isInitialized()) { 3989 memoizedIsInitialized = 0; 3990 return false; 3991 } 3992 if (hasDestServerName()) { 3993 if (!getDestServerName().isInitialized()) { 3994 memoizedIsInitialized = 0; 3995 return false; 3996 } 3997 } 3998 memoizedIsInitialized = 1; 3999 return true; 4000 } 4001 writeTo(com.google.protobuf.CodedOutputStream output)4002 public void writeTo(com.google.protobuf.CodedOutputStream output) 4003 throws java.io.IOException { 4004 getSerializedSize(); 4005 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4006 output.writeMessage(1, region_); 4007 } 4008 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4009 output.writeMessage(2, destServerName_); 4010 } 4011 getUnknownFields().writeTo(output); 4012 } 4013 4014 private int memoizedSerializedSize = -1; getSerializedSize()4015 public int getSerializedSize() { 4016 int size = memoizedSerializedSize; 4017 if (size != -1) return size; 4018 4019 size = 0; 4020 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4021 size += com.google.protobuf.CodedOutputStream 4022 .computeMessageSize(1, region_); 4023 } 4024 if (((bitField0_ & 0x00000002) == 0x00000002)) { 4025 size += com.google.protobuf.CodedOutputStream 4026 .computeMessageSize(2, destServerName_); 4027 } 4028 size += getUnknownFields().getSerializedSize(); 4029 memoizedSerializedSize = size; 4030 return size; 4031 } 4032 4033 private static final long serialVersionUID = 0L; 4034 @java.lang.Override writeReplace()4035 protected java.lang.Object writeReplace() 4036 throws java.io.ObjectStreamException { 4037 return super.writeReplace(); 4038 } 4039 4040 @java.lang.Override equals(final java.lang.Object obj)4041 public boolean equals(final java.lang.Object obj) { 4042 if (obj == this) { 4043 return true; 4044 } 4045 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest)) { 4046 return super.equals(obj); 4047 } 4048 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest) obj; 4049 4050 boolean result = true; 4051 result = result && (hasRegion() == other.hasRegion()); 4052 if (hasRegion()) { 4053 result = result && getRegion() 4054 .equals(other.getRegion()); 4055 } 4056 result = result && (hasDestServerName() == other.hasDestServerName()); 4057 if (hasDestServerName()) { 4058 result = result && getDestServerName() 4059 .equals(other.getDestServerName()); 4060 } 4061 result = result && 4062 getUnknownFields().equals(other.getUnknownFields()); 4063 return result; 4064 } 4065 4066 private int memoizedHashCode = 0; 4067 @java.lang.Override hashCode()4068 public int hashCode() { 4069 if (memoizedHashCode != 0) { 4070 return memoizedHashCode; 4071 } 4072 int hash = 41; 4073 hash = (19 * hash) + getDescriptorForType().hashCode(); 4074 if (hasRegion()) { 4075 hash = (37 * hash) + REGION_FIELD_NUMBER; 4076 hash = (53 * hash) + getRegion().hashCode(); 4077 } 4078 if (hasDestServerName()) { 4079 hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER; 4080 hash = (53 * hash) + getDestServerName().hashCode(); 4081 } 4082 hash = (29 * hash) + getUnknownFields().hashCode(); 4083 memoizedHashCode = hash; 4084 return hash; 4085 } 4086 parseFrom( com.google.protobuf.ByteString data)4087 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4088 com.google.protobuf.ByteString data) 4089 throws com.google.protobuf.InvalidProtocolBufferException { 4090 return PARSER.parseFrom(data); 4091 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4092 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4093 com.google.protobuf.ByteString data, 4094 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4095 throws com.google.protobuf.InvalidProtocolBufferException { 4096 return PARSER.parseFrom(data, extensionRegistry); 4097 } parseFrom(byte[] data)4098 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(byte[] data) 4099 throws com.google.protobuf.InvalidProtocolBufferException { 4100 return PARSER.parseFrom(data); 4101 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4102 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4103 byte[] data, 4104 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4105 throws com.google.protobuf.InvalidProtocolBufferException { 4106 return PARSER.parseFrom(data, extensionRegistry); 4107 } parseFrom(java.io.InputStream input)4108 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(java.io.InputStream input) 4109 throws java.io.IOException { 4110 return PARSER.parseFrom(input); 4111 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4112 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4113 java.io.InputStream input, 4114 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4115 throws java.io.IOException { 4116 return PARSER.parseFrom(input, extensionRegistry); 4117 } parseDelimitedFrom(java.io.InputStream input)4118 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseDelimitedFrom(java.io.InputStream input) 4119 throws java.io.IOException { 4120 return PARSER.parseDelimitedFrom(input); 4121 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4122 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseDelimitedFrom( 4123 java.io.InputStream input, 4124 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4125 throws java.io.IOException { 4126 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4127 } parseFrom( com.google.protobuf.CodedInputStream input)4128 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4129 com.google.protobuf.CodedInputStream input) 4130 throws java.io.IOException { 4131 return PARSER.parseFrom(input); 4132 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4133 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom( 4134 com.google.protobuf.CodedInputStream input, 4135 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4136 throws java.io.IOException { 4137 return PARSER.parseFrom(input, extensionRegistry); 4138 } 4139 newBuilder()4140 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()4141 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest prototype)4142 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest prototype) { 4143 return newBuilder().mergeFrom(prototype); 4144 } toBuilder()4145 public Builder toBuilder() { return newBuilder(this); } 4146 4147 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)4148 protected Builder newBuilderForType( 4149 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4150 Builder builder = new Builder(parent); 4151 return builder; 4152 } 4153 /** 4154 * Protobuf type {@code MoveRegionRequest} 4155 */ 4156 public static final class Builder extends 4157 com.google.protobuf.GeneratedMessage.Builder<Builder> 4158 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequestOrBuilder { 4159 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()4160 getDescriptor() { 4161 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionRequest_descriptor; 4162 } 4163 4164 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()4165 internalGetFieldAccessorTable() { 4166 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionRequest_fieldAccessorTable 4167 .ensureFieldAccessorsInitialized( 4168 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.Builder.class); 4169 } 4170 4171 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.newBuilder() Builder()4172 private Builder() { 4173 maybeForceBuilderInitialization(); 4174 } 4175 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)4176 private Builder( 4177 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4178 super(parent); 4179 maybeForceBuilderInitialization(); 4180 } maybeForceBuilderInitialization()4181 private void maybeForceBuilderInitialization() { 4182 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4183 getRegionFieldBuilder(); 4184 getDestServerNameFieldBuilder(); 4185 } 4186 } create()4187 private static Builder create() { 4188 return new Builder(); 4189 } 4190 clear()4191 public Builder clear() { 4192 super.clear(); 4193 if (regionBuilder_ == null) { 4194 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 4195 } else { 4196 regionBuilder_.clear(); 4197 } 4198 bitField0_ = (bitField0_ & ~0x00000001); 4199 if (destServerNameBuilder_ == null) { 4200 destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); 4201 } else { 4202 destServerNameBuilder_.clear(); 4203 } 4204 bitField0_ = (bitField0_ & ~0x00000002); 4205 return this; 4206 } 4207 clone()4208 public Builder clone() { 4209 return create().mergeFrom(buildPartial()); 4210 } 4211 4212 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()4213 getDescriptorForType() { 4214 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionRequest_descriptor; 4215 } 4216 getDefaultInstanceForType()4217 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest getDefaultInstanceForType() { 4218 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance(); 4219 } 4220 build()4221 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest build() { 4222 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest result = buildPartial(); 4223 if (!result.isInitialized()) { 4224 throw newUninitializedMessageException(result); 4225 } 4226 return result; 4227 } 4228 buildPartial()4229 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest buildPartial() { 4230 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest(this); 4231 int from_bitField0_ = bitField0_; 4232 int to_bitField0_ = 0; 4233 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 4234 to_bitField0_ |= 0x00000001; 4235 } 4236 if (regionBuilder_ == null) { 4237 result.region_ = region_; 4238 } else { 4239 result.region_ = regionBuilder_.build(); 4240 } 4241 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 4242 to_bitField0_ |= 0x00000002; 4243 } 4244 if (destServerNameBuilder_ == null) { 4245 result.destServerName_ = destServerName_; 4246 } else { 4247 result.destServerName_ = destServerNameBuilder_.build(); 4248 } 4249 result.bitField0_ = to_bitField0_; 4250 onBuilt(); 4251 return result; 4252 } 4253 mergeFrom(com.google.protobuf.Message other)4254 public Builder mergeFrom(com.google.protobuf.Message other) { 4255 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest) { 4256 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest)other); 4257 } else { 4258 super.mergeFrom(other); 4259 return this; 4260 } 4261 } 4262 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest other)4263 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest other) { 4264 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance()) return this; 4265 if (other.hasRegion()) { 4266 mergeRegion(other.getRegion()); 4267 } 4268 if (other.hasDestServerName()) { 4269 mergeDestServerName(other.getDestServerName()); 4270 } 4271 this.mergeUnknownFields(other.getUnknownFields()); 4272 return this; 4273 } 4274 isInitialized()4275 public final boolean isInitialized() { 4276 if (!hasRegion()) { 4277 4278 return false; 4279 } 4280 if (!getRegion().isInitialized()) { 4281 4282 return false; 4283 } 4284 if (hasDestServerName()) { 4285 if (!getDestServerName().isInitialized()) { 4286 4287 return false; 4288 } 4289 } 4290 return true; 4291 } 4292 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4293 public Builder mergeFrom( 4294 com.google.protobuf.CodedInputStream input, 4295 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4296 throws java.io.IOException { 4297 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest parsedMessage = null; 4298 try { 4299 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4300 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4301 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionRequest) e.getUnfinishedMessage(); 4302 throw e; 4303 } finally { 4304 if (parsedMessage != null) { 4305 mergeFrom(parsedMessage); 4306 } 4307 } 4308 return this; 4309 } 4310 private int bitField0_; 4311 4312 // required .RegionSpecifier region = 1; 4313 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 4314 private com.google.protobuf.SingleFieldBuilder< 4315 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; 4316 /** 4317 * <code>required .RegionSpecifier region = 1;</code> 4318 */ hasRegion()4319 public boolean hasRegion() { 4320 return ((bitField0_ & 0x00000001) == 0x00000001); 4321 } 4322 /** 4323 * <code>required .RegionSpecifier region = 1;</code> 4324 */ getRegion()4325 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 4326 if (regionBuilder_ == null) { 4327 return region_; 4328 } else { 4329 return regionBuilder_.getMessage(); 4330 } 4331 } 4332 /** 4333 * <code>required .RegionSpecifier region = 1;</code> 4334 */ setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)4335 public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 4336 if (regionBuilder_ == null) { 4337 if (value == null) { 4338 throw new NullPointerException(); 4339 } 4340 region_ = value; 4341 onChanged(); 4342 } else { 4343 regionBuilder_.setMessage(value); 4344 } 4345 bitField0_ |= 0x00000001; 4346 return this; 4347 } 4348 /** 4349 * <code>required .RegionSpecifier region = 1;</code> 4350 */ setRegion( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)4351 public Builder setRegion( 4352 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 4353 if (regionBuilder_ == null) { 4354 region_ = builderForValue.build(); 4355 onChanged(); 4356 } else { 4357 regionBuilder_.setMessage(builderForValue.build()); 4358 } 4359 bitField0_ |= 0x00000001; 4360 return this; 4361 } 4362 /** 4363 * <code>required .RegionSpecifier region = 1;</code> 4364 */ mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)4365 public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 4366 if (regionBuilder_ == null) { 4367 if (((bitField0_ & 0x00000001) == 0x00000001) && 4368 region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 4369 region_ = 4370 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); 4371 } else { 4372 region_ = value; 4373 } 4374 onChanged(); 4375 } else { 4376 regionBuilder_.mergeFrom(value); 4377 } 4378 bitField0_ |= 0x00000001; 4379 return this; 4380 } 4381 /** 4382 * <code>required .RegionSpecifier region = 1;</code> 4383 */ clearRegion()4384 public Builder clearRegion() { 4385 if (regionBuilder_ == null) { 4386 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 4387 onChanged(); 4388 } else { 4389 regionBuilder_.clear(); 4390 } 4391 bitField0_ = (bitField0_ & ~0x00000001); 4392 return this; 4393 } 4394 /** 4395 * <code>required .RegionSpecifier region = 1;</code> 4396 */ getRegionBuilder()4397 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { 4398 bitField0_ |= 0x00000001; 4399 onChanged(); 4400 return getRegionFieldBuilder().getBuilder(); 4401 } 4402 /** 4403 * <code>required .RegionSpecifier region = 1;</code> 4404 */ getRegionOrBuilder()4405 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 4406 if (regionBuilder_ != null) { 4407 return regionBuilder_.getMessageOrBuilder(); 4408 } else { 4409 return region_; 4410 } 4411 } 4412 /** 4413 * <code>required .RegionSpecifier region = 1;</code> 4414 */ 4415 private com.google.protobuf.SingleFieldBuilder< 4416 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder()4417 getRegionFieldBuilder() { 4418 if (regionBuilder_ == null) { 4419 regionBuilder_ = new com.google.protobuf.SingleFieldBuilder< 4420 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 4421 region_, 4422 getParentForChildren(), 4423 isClean()); 4424 region_ = null; 4425 } 4426 return regionBuilder_; 4427 } 4428 4429 // optional .ServerName dest_server_name = 2; 4430 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); 4431 private com.google.protobuf.SingleFieldBuilder< 4432 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_; 4433 /** 4434 * <code>optional .ServerName dest_server_name = 2;</code> 4435 */ hasDestServerName()4436 public boolean hasDestServerName() { 4437 return ((bitField0_ & 0x00000002) == 0x00000002); 4438 } 4439 /** 4440 * <code>optional .ServerName dest_server_name = 2;</code> 4441 */ getDestServerName()4442 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getDestServerName() { 4443 if (destServerNameBuilder_ == null) { 4444 return destServerName_; 4445 } else { 4446 return destServerNameBuilder_.getMessage(); 4447 } 4448 } 4449 /** 4450 * <code>optional .ServerName dest_server_name = 2;</code> 4451 */ setDestServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value)4452 public Builder setDestServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { 4453 if (destServerNameBuilder_ == null) { 4454 if (value == null) { 4455 throw new NullPointerException(); 4456 } 4457 destServerName_ = value; 4458 onChanged(); 4459 } else { 4460 destServerNameBuilder_.setMessage(value); 4461 } 4462 bitField0_ |= 0x00000002; 4463 return this; 4464 } 4465 /** 4466 * <code>optional .ServerName dest_server_name = 2;</code> 4467 */ setDestServerName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue)4468 public Builder setDestServerName( 4469 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { 4470 if (destServerNameBuilder_ == null) { 4471 destServerName_ = builderForValue.build(); 4472 onChanged(); 4473 } else { 4474 destServerNameBuilder_.setMessage(builderForValue.build()); 4475 } 4476 bitField0_ |= 0x00000002; 4477 return this; 4478 } 4479 /** 4480 * <code>optional .ServerName dest_server_name = 2;</code> 4481 */ mergeDestServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value)4482 public Builder mergeDestServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { 4483 if (destServerNameBuilder_ == null) { 4484 if (((bitField0_ & 0x00000002) == 0x00000002) && 4485 destServerName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { 4486 destServerName_ = 4487 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(destServerName_).mergeFrom(value).buildPartial(); 4488 } else { 4489 destServerName_ = value; 4490 } 4491 onChanged(); 4492 } else { 4493 destServerNameBuilder_.mergeFrom(value); 4494 } 4495 bitField0_ |= 0x00000002; 4496 return this; 4497 } 4498 /** 4499 * <code>optional .ServerName dest_server_name = 2;</code> 4500 */ clearDestServerName()4501 public Builder clearDestServerName() { 4502 if (destServerNameBuilder_ == null) { 4503 destServerName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); 4504 onChanged(); 4505 } else { 4506 destServerNameBuilder_.clear(); 4507 } 4508 bitField0_ = (bitField0_ & ~0x00000002); 4509 return this; 4510 } 4511 /** 4512 * <code>optional .ServerName dest_server_name = 2;</code> 4513 */ getDestServerNameBuilder()4514 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getDestServerNameBuilder() { 4515 bitField0_ |= 0x00000002; 4516 onChanged(); 4517 return getDestServerNameFieldBuilder().getBuilder(); 4518 } 4519 /** 4520 * <code>optional .ServerName dest_server_name = 2;</code> 4521 */ getDestServerNameOrBuilder()4522 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder() { 4523 if (destServerNameBuilder_ != null) { 4524 return destServerNameBuilder_.getMessageOrBuilder(); 4525 } else { 4526 return destServerName_; 4527 } 4528 } 4529 /** 4530 * <code>optional .ServerName dest_server_name = 2;</code> 4531 */ 4532 private com.google.protobuf.SingleFieldBuilder< 4533 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getDestServerNameFieldBuilder()4534 getDestServerNameFieldBuilder() { 4535 if (destServerNameBuilder_ == null) { 4536 destServerNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< 4537 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( 4538 destServerName_, 4539 getParentForChildren(), 4540 isClean()); 4541 destServerName_ = null; 4542 } 4543 return destServerNameBuilder_; 4544 } 4545 4546 // @@protoc_insertion_point(builder_scope:MoveRegionRequest) 4547 } 4548 4549 static { 4550 defaultInstance = new MoveRegionRequest(true); defaultInstance.initFields()4551 defaultInstance.initFields(); 4552 } 4553 4554 // @@protoc_insertion_point(class_scope:MoveRegionRequest) 4555 } 4556 4557 public interface MoveRegionResponseOrBuilder 4558 extends com.google.protobuf.MessageOrBuilder { 4559 } 4560 /** 4561 * Protobuf type {@code MoveRegionResponse} 4562 */ 4563 public static final class MoveRegionResponse extends 4564 com.google.protobuf.GeneratedMessage 4565 implements MoveRegionResponseOrBuilder { 4566 // Use MoveRegionResponse.newBuilder() to construct. MoveRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)4567 private MoveRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4568 super(builder); 4569 this.unknownFields = builder.getUnknownFields(); 4570 } MoveRegionResponse(boolean noInit)4571 private MoveRegionResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4572 4573 private static final MoveRegionResponse defaultInstance; getDefaultInstance()4574 public static MoveRegionResponse getDefaultInstance() { 4575 return defaultInstance; 4576 } 4577 getDefaultInstanceForType()4578 public MoveRegionResponse getDefaultInstanceForType() { 4579 return defaultInstance; 4580 } 4581 4582 private final com.google.protobuf.UnknownFieldSet unknownFields; 4583 @java.lang.Override 4584 public final com.google.protobuf.UnknownFieldSet getUnknownFields()4585 getUnknownFields() { 4586 return this.unknownFields; 4587 } MoveRegionResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4588 private MoveRegionResponse( 4589 com.google.protobuf.CodedInputStream input, 4590 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4591 throws com.google.protobuf.InvalidProtocolBufferException { 4592 initFields(); 4593 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4594 com.google.protobuf.UnknownFieldSet.newBuilder(); 4595 try { 4596 boolean done = false; 4597 while (!done) { 4598 int tag = input.readTag(); 4599 switch (tag) { 4600 case 0: 4601 done = true; 4602 break; 4603 default: { 4604 if (!parseUnknownField(input, unknownFields, 4605 extensionRegistry, tag)) { 4606 done = true; 4607 } 4608 break; 4609 } 4610 } 4611 } 4612 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4613 throw e.setUnfinishedMessage(this); 4614 } catch (java.io.IOException e) { 4615 throw new com.google.protobuf.InvalidProtocolBufferException( 4616 e.getMessage()).setUnfinishedMessage(this); 4617 } finally { 4618 this.unknownFields = unknownFields.build(); 4619 makeExtensionsImmutable(); 4620 } 4621 } 4622 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()4623 getDescriptor() { 4624 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionResponse_descriptor; 4625 } 4626 4627 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()4628 internalGetFieldAccessorTable() { 4629 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionResponse_fieldAccessorTable 4630 .ensureFieldAccessorsInitialized( 4631 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.Builder.class); 4632 } 4633 4634 public static com.google.protobuf.Parser<MoveRegionResponse> PARSER = 4635 new com.google.protobuf.AbstractParser<MoveRegionResponse>() { 4636 public MoveRegionResponse parsePartialFrom( 4637 com.google.protobuf.CodedInputStream input, 4638 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4639 throws com.google.protobuf.InvalidProtocolBufferException { 4640 return new MoveRegionResponse(input, extensionRegistry); 4641 } 4642 }; 4643 4644 @java.lang.Override getParserForType()4645 public com.google.protobuf.Parser<MoveRegionResponse> getParserForType() { 4646 return PARSER; 4647 } 4648 initFields()4649 private void initFields() { 4650 } 4651 private byte memoizedIsInitialized = -1; isInitialized()4652 public final boolean isInitialized() { 4653 byte isInitialized = memoizedIsInitialized; 4654 if (isInitialized != -1) return isInitialized == 1; 4655 4656 memoizedIsInitialized = 1; 4657 return true; 4658 } 4659 writeTo(com.google.protobuf.CodedOutputStream output)4660 public void writeTo(com.google.protobuf.CodedOutputStream output) 4661 throws java.io.IOException { 4662 getSerializedSize(); 4663 getUnknownFields().writeTo(output); 4664 } 4665 4666 private int memoizedSerializedSize = -1; getSerializedSize()4667 public int getSerializedSize() { 4668 int size = memoizedSerializedSize; 4669 if (size != -1) return size; 4670 4671 size = 0; 4672 size += getUnknownFields().getSerializedSize(); 4673 memoizedSerializedSize = size; 4674 return size; 4675 } 4676 4677 private static final long serialVersionUID = 0L; 4678 @java.lang.Override writeReplace()4679 protected java.lang.Object writeReplace() 4680 throws java.io.ObjectStreamException { 4681 return super.writeReplace(); 4682 } 4683 4684 @java.lang.Override equals(final java.lang.Object obj)4685 public boolean equals(final java.lang.Object obj) { 4686 if (obj == this) { 4687 return true; 4688 } 4689 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse)) { 4690 return super.equals(obj); 4691 } 4692 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse) obj; 4693 4694 boolean result = true; 4695 result = result && 4696 getUnknownFields().equals(other.getUnknownFields()); 4697 return result; 4698 } 4699 4700 private int memoizedHashCode = 0; 4701 @java.lang.Override hashCode()4702 public int hashCode() { 4703 if (memoizedHashCode != 0) { 4704 return memoizedHashCode; 4705 } 4706 int hash = 41; 4707 hash = (19 * hash) + getDescriptorForType().hashCode(); 4708 hash = (29 * hash) + getUnknownFields().hashCode(); 4709 memoizedHashCode = hash; 4710 return hash; 4711 } 4712 parseFrom( com.google.protobuf.ByteString data)4713 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4714 com.google.protobuf.ByteString data) 4715 throws com.google.protobuf.InvalidProtocolBufferException { 4716 return PARSER.parseFrom(data); 4717 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4718 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4719 com.google.protobuf.ByteString data, 4720 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4721 throws com.google.protobuf.InvalidProtocolBufferException { 4722 return PARSER.parseFrom(data, extensionRegistry); 4723 } parseFrom(byte[] data)4724 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(byte[] data) 4725 throws com.google.protobuf.InvalidProtocolBufferException { 4726 return PARSER.parseFrom(data); 4727 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4728 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4729 byte[] data, 4730 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4731 throws com.google.protobuf.InvalidProtocolBufferException { 4732 return PARSER.parseFrom(data, extensionRegistry); 4733 } parseFrom(java.io.InputStream input)4734 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(java.io.InputStream input) 4735 throws java.io.IOException { 4736 return PARSER.parseFrom(input); 4737 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4738 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4739 java.io.InputStream input, 4740 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4741 throws java.io.IOException { 4742 return PARSER.parseFrom(input, extensionRegistry); 4743 } parseDelimitedFrom(java.io.InputStream input)4744 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseDelimitedFrom(java.io.InputStream input) 4745 throws java.io.IOException { 4746 return PARSER.parseDelimitedFrom(input); 4747 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4748 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseDelimitedFrom( 4749 java.io.InputStream input, 4750 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4751 throws java.io.IOException { 4752 return PARSER.parseDelimitedFrom(input, extensionRegistry); 4753 } parseFrom( com.google.protobuf.CodedInputStream input)4754 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4755 com.google.protobuf.CodedInputStream input) 4756 throws java.io.IOException { 4757 return PARSER.parseFrom(input); 4758 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4759 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom( 4760 com.google.protobuf.CodedInputStream input, 4761 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4762 throws java.io.IOException { 4763 return PARSER.parseFrom(input, extensionRegistry); 4764 } 4765 newBuilder()4766 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()4767 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse prototype)4768 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse prototype) { 4769 return newBuilder().mergeFrom(prototype); 4770 } toBuilder()4771 public Builder toBuilder() { return newBuilder(this); } 4772 4773 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)4774 protected Builder newBuilderForType( 4775 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4776 Builder builder = new Builder(parent); 4777 return builder; 4778 } 4779 /** 4780 * Protobuf type {@code MoveRegionResponse} 4781 */ 4782 public static final class Builder extends 4783 com.google.protobuf.GeneratedMessage.Builder<Builder> 4784 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponseOrBuilder { 4785 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()4786 getDescriptor() { 4787 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionResponse_descriptor; 4788 } 4789 4790 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()4791 internalGetFieldAccessorTable() { 4792 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionResponse_fieldAccessorTable 4793 .ensureFieldAccessorsInitialized( 4794 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.Builder.class); 4795 } 4796 4797 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.newBuilder() Builder()4798 private Builder() { 4799 maybeForceBuilderInitialization(); 4800 } 4801 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)4802 private Builder( 4803 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4804 super(parent); 4805 maybeForceBuilderInitialization(); 4806 } maybeForceBuilderInitialization()4807 private void maybeForceBuilderInitialization() { 4808 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4809 } 4810 } create()4811 private static Builder create() { 4812 return new Builder(); 4813 } 4814 clear()4815 public Builder clear() { 4816 super.clear(); 4817 return this; 4818 } 4819 clone()4820 public Builder clone() { 4821 return create().mergeFrom(buildPartial()); 4822 } 4823 4824 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()4825 getDescriptorForType() { 4826 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_MoveRegionResponse_descriptor; 4827 } 4828 getDefaultInstanceForType()4829 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse getDefaultInstanceForType() { 4830 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(); 4831 } 4832 build()4833 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse build() { 4834 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse result = buildPartial(); 4835 if (!result.isInitialized()) { 4836 throw newUninitializedMessageException(result); 4837 } 4838 return result; 4839 } 4840 buildPartial()4841 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse buildPartial() { 4842 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse(this); 4843 onBuilt(); 4844 return result; 4845 } 4846 mergeFrom(com.google.protobuf.Message other)4847 public Builder mergeFrom(com.google.protobuf.Message other) { 4848 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse) { 4849 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse)other); 4850 } else { 4851 super.mergeFrom(other); 4852 return this; 4853 } 4854 } 4855 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse other)4856 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse other) { 4857 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance()) return this; 4858 this.mergeUnknownFields(other.getUnknownFields()); 4859 return this; 4860 } 4861 isInitialized()4862 public final boolean isInitialized() { 4863 return true; 4864 } 4865 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4866 public Builder mergeFrom( 4867 com.google.protobuf.CodedInputStream input, 4868 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4869 throws java.io.IOException { 4870 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse parsedMessage = null; 4871 try { 4872 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4873 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4874 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MoveRegionResponse) e.getUnfinishedMessage(); 4875 throw e; 4876 } finally { 4877 if (parsedMessage != null) { 4878 mergeFrom(parsedMessage); 4879 } 4880 } 4881 return this; 4882 } 4883 4884 // @@protoc_insertion_point(builder_scope:MoveRegionResponse) 4885 } 4886 4887 static { 4888 defaultInstance = new MoveRegionResponse(true); defaultInstance.initFields()4889 defaultInstance.initFields(); 4890 } 4891 4892 // @@protoc_insertion_point(class_scope:MoveRegionResponse) 4893 } 4894 4895 public interface DispatchMergingRegionsRequestOrBuilder 4896 extends com.google.protobuf.MessageOrBuilder { 4897 4898 // required .RegionSpecifier region_a = 1; 4899 /** 4900 * <code>required .RegionSpecifier region_a = 1;</code> 4901 */ hasRegionA()4902 boolean hasRegionA(); 4903 /** 4904 * <code>required .RegionSpecifier region_a = 1;</code> 4905 */ getRegionA()4906 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA(); 4907 /** 4908 * <code>required .RegionSpecifier region_a = 1;</code> 4909 */ getRegionAOrBuilder()4910 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder(); 4911 4912 // required .RegionSpecifier region_b = 2; 4913 /** 4914 * <code>required .RegionSpecifier region_b = 2;</code> 4915 */ hasRegionB()4916 boolean hasRegionB(); 4917 /** 4918 * <code>required .RegionSpecifier region_b = 2;</code> 4919 */ getRegionB()4920 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB(); 4921 /** 4922 * <code>required .RegionSpecifier region_b = 2;</code> 4923 */ getRegionBOrBuilder()4924 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder(); 4925 4926 // optional bool forcible = 3 [default = false]; 4927 /** 4928 * <code>optional bool forcible = 3 [default = false];</code> 4929 */ hasForcible()4930 boolean hasForcible(); 4931 /** 4932 * <code>optional bool forcible = 3 [default = false];</code> 4933 */ getForcible()4934 boolean getForcible(); 4935 } 4936 /** 4937 * Protobuf type {@code DispatchMergingRegionsRequest} 4938 * 4939 * <pre> 4940 ** 4941 * Dispatch merging the specified regions. 4942 * </pre> 4943 */ 4944 public static final class DispatchMergingRegionsRequest extends 4945 com.google.protobuf.GeneratedMessage 4946 implements DispatchMergingRegionsRequestOrBuilder { 4947 // Use DispatchMergingRegionsRequest.newBuilder() to construct. DispatchMergingRegionsRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)4948 private DispatchMergingRegionsRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 4949 super(builder); 4950 this.unknownFields = builder.getUnknownFields(); 4951 } DispatchMergingRegionsRequest(boolean noInit)4952 private DispatchMergingRegionsRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 4953 4954 private static final DispatchMergingRegionsRequest defaultInstance; getDefaultInstance()4955 public static DispatchMergingRegionsRequest getDefaultInstance() { 4956 return defaultInstance; 4957 } 4958 getDefaultInstanceForType()4959 public DispatchMergingRegionsRequest getDefaultInstanceForType() { 4960 return defaultInstance; 4961 } 4962 4963 private final com.google.protobuf.UnknownFieldSet unknownFields; 4964 @java.lang.Override 4965 public final com.google.protobuf.UnknownFieldSet getUnknownFields()4966 getUnknownFields() { 4967 return this.unknownFields; 4968 } DispatchMergingRegionsRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)4969 private DispatchMergingRegionsRequest( 4970 com.google.protobuf.CodedInputStream input, 4971 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4972 throws com.google.protobuf.InvalidProtocolBufferException { 4973 initFields(); 4974 int mutable_bitField0_ = 0; 4975 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 4976 com.google.protobuf.UnknownFieldSet.newBuilder(); 4977 try { 4978 boolean done = false; 4979 while (!done) { 4980 int tag = input.readTag(); 4981 switch (tag) { 4982 case 0: 4983 done = true; 4984 break; 4985 default: { 4986 if (!parseUnknownField(input, unknownFields, 4987 extensionRegistry, tag)) { 4988 done = true; 4989 } 4990 break; 4991 } 4992 case 10: { 4993 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 4994 if (((bitField0_ & 0x00000001) == 0x00000001)) { 4995 subBuilder = regionA_.toBuilder(); 4996 } 4997 regionA_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 4998 if (subBuilder != null) { 4999 subBuilder.mergeFrom(regionA_); 5000 regionA_ = subBuilder.buildPartial(); 5001 } 5002 bitField0_ |= 0x00000001; 5003 break; 5004 } 5005 case 18: { 5006 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 5007 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5008 subBuilder = regionB_.toBuilder(); 5009 } 5010 regionB_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 5011 if (subBuilder != null) { 5012 subBuilder.mergeFrom(regionB_); 5013 regionB_ = subBuilder.buildPartial(); 5014 } 5015 bitField0_ |= 0x00000002; 5016 break; 5017 } 5018 case 24: { 5019 bitField0_ |= 0x00000004; 5020 forcible_ = input.readBool(); 5021 break; 5022 } 5023 } 5024 } 5025 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5026 throw e.setUnfinishedMessage(this); 5027 } catch (java.io.IOException e) { 5028 throw new com.google.protobuf.InvalidProtocolBufferException( 5029 e.getMessage()).setUnfinishedMessage(this); 5030 } finally { 5031 this.unknownFields = unknownFields.build(); 5032 makeExtensionsImmutable(); 5033 } 5034 } 5035 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()5036 getDescriptor() { 5037 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsRequest_descriptor; 5038 } 5039 5040 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()5041 internalGetFieldAccessorTable() { 5042 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsRequest_fieldAccessorTable 5043 .ensureFieldAccessorsInitialized( 5044 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class); 5045 } 5046 5047 public static com.google.protobuf.Parser<DispatchMergingRegionsRequest> PARSER = 5048 new com.google.protobuf.AbstractParser<DispatchMergingRegionsRequest>() { 5049 public DispatchMergingRegionsRequest parsePartialFrom( 5050 com.google.protobuf.CodedInputStream input, 5051 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5052 throws com.google.protobuf.InvalidProtocolBufferException { 5053 return new DispatchMergingRegionsRequest(input, extensionRegistry); 5054 } 5055 }; 5056 5057 @java.lang.Override getParserForType()5058 public com.google.protobuf.Parser<DispatchMergingRegionsRequest> getParserForType() { 5059 return PARSER; 5060 } 5061 5062 private int bitField0_; 5063 // required .RegionSpecifier region_a = 1; 5064 public static final int REGION_A_FIELD_NUMBER = 1; 5065 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_; 5066 /** 5067 * <code>required .RegionSpecifier region_a = 1;</code> 5068 */ hasRegionA()5069 public boolean hasRegionA() { 5070 return ((bitField0_ & 0x00000001) == 0x00000001); 5071 } 5072 /** 5073 * <code>required .RegionSpecifier region_a = 1;</code> 5074 */ getRegionA()5075 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { 5076 return regionA_; 5077 } 5078 /** 5079 * <code>required .RegionSpecifier region_a = 1;</code> 5080 */ getRegionAOrBuilder()5081 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { 5082 return regionA_; 5083 } 5084 5085 // required .RegionSpecifier region_b = 2; 5086 public static final int REGION_B_FIELD_NUMBER = 2; 5087 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_; 5088 /** 5089 * <code>required .RegionSpecifier region_b = 2;</code> 5090 */ hasRegionB()5091 public boolean hasRegionB() { 5092 return ((bitField0_ & 0x00000002) == 0x00000002); 5093 } 5094 /** 5095 * <code>required .RegionSpecifier region_b = 2;</code> 5096 */ getRegionB()5097 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { 5098 return regionB_; 5099 } 5100 /** 5101 * <code>required .RegionSpecifier region_b = 2;</code> 5102 */ getRegionBOrBuilder()5103 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { 5104 return regionB_; 5105 } 5106 5107 // optional bool forcible = 3 [default = false]; 5108 public static final int FORCIBLE_FIELD_NUMBER = 3; 5109 private boolean forcible_; 5110 /** 5111 * <code>optional bool forcible = 3 [default = false];</code> 5112 */ hasForcible()5113 public boolean hasForcible() { 5114 return ((bitField0_ & 0x00000004) == 0x00000004); 5115 } 5116 /** 5117 * <code>optional bool forcible = 3 [default = false];</code> 5118 */ getForcible()5119 public boolean getForcible() { 5120 return forcible_; 5121 } 5122 initFields()5123 private void initFields() { 5124 regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5125 regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5126 forcible_ = false; 5127 } 5128 private byte memoizedIsInitialized = -1; isInitialized()5129 public final boolean isInitialized() { 5130 byte isInitialized = memoizedIsInitialized; 5131 if (isInitialized != -1) return isInitialized == 1; 5132 5133 if (!hasRegionA()) { 5134 memoizedIsInitialized = 0; 5135 return false; 5136 } 5137 if (!hasRegionB()) { 5138 memoizedIsInitialized = 0; 5139 return false; 5140 } 5141 if (!getRegionA().isInitialized()) { 5142 memoizedIsInitialized = 0; 5143 return false; 5144 } 5145 if (!getRegionB().isInitialized()) { 5146 memoizedIsInitialized = 0; 5147 return false; 5148 } 5149 memoizedIsInitialized = 1; 5150 return true; 5151 } 5152 writeTo(com.google.protobuf.CodedOutputStream output)5153 public void writeTo(com.google.protobuf.CodedOutputStream output) 5154 throws java.io.IOException { 5155 getSerializedSize(); 5156 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5157 output.writeMessage(1, regionA_); 5158 } 5159 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5160 output.writeMessage(2, regionB_); 5161 } 5162 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5163 output.writeBool(3, forcible_); 5164 } 5165 getUnknownFields().writeTo(output); 5166 } 5167 5168 private int memoizedSerializedSize = -1; getSerializedSize()5169 public int getSerializedSize() { 5170 int size = memoizedSerializedSize; 5171 if (size != -1) return size; 5172 5173 size = 0; 5174 if (((bitField0_ & 0x00000001) == 0x00000001)) { 5175 size += com.google.protobuf.CodedOutputStream 5176 .computeMessageSize(1, regionA_); 5177 } 5178 if (((bitField0_ & 0x00000002) == 0x00000002)) { 5179 size += com.google.protobuf.CodedOutputStream 5180 .computeMessageSize(2, regionB_); 5181 } 5182 if (((bitField0_ & 0x00000004) == 0x00000004)) { 5183 size += com.google.protobuf.CodedOutputStream 5184 .computeBoolSize(3, forcible_); 5185 } 5186 size += getUnknownFields().getSerializedSize(); 5187 memoizedSerializedSize = size; 5188 return size; 5189 } 5190 5191 private static final long serialVersionUID = 0L; 5192 @java.lang.Override writeReplace()5193 protected java.lang.Object writeReplace() 5194 throws java.io.ObjectStreamException { 5195 return super.writeReplace(); 5196 } 5197 5198 @java.lang.Override equals(final java.lang.Object obj)5199 public boolean equals(final java.lang.Object obj) { 5200 if (obj == this) { 5201 return true; 5202 } 5203 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)) { 5204 return super.equals(obj); 5205 } 5206 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) obj; 5207 5208 boolean result = true; 5209 result = result && (hasRegionA() == other.hasRegionA()); 5210 if (hasRegionA()) { 5211 result = result && getRegionA() 5212 .equals(other.getRegionA()); 5213 } 5214 result = result && (hasRegionB() == other.hasRegionB()); 5215 if (hasRegionB()) { 5216 result = result && getRegionB() 5217 .equals(other.getRegionB()); 5218 } 5219 result = result && (hasForcible() == other.hasForcible()); 5220 if (hasForcible()) { 5221 result = result && (getForcible() 5222 == other.getForcible()); 5223 } 5224 result = result && 5225 getUnknownFields().equals(other.getUnknownFields()); 5226 return result; 5227 } 5228 5229 private int memoizedHashCode = 0; 5230 @java.lang.Override hashCode()5231 public int hashCode() { 5232 if (memoizedHashCode != 0) { 5233 return memoizedHashCode; 5234 } 5235 int hash = 41; 5236 hash = (19 * hash) + getDescriptorForType().hashCode(); 5237 if (hasRegionA()) { 5238 hash = (37 * hash) + REGION_A_FIELD_NUMBER; 5239 hash = (53 * hash) + getRegionA().hashCode(); 5240 } 5241 if (hasRegionB()) { 5242 hash = (37 * hash) + REGION_B_FIELD_NUMBER; 5243 hash = (53 * hash) + getRegionB().hashCode(); 5244 } 5245 if (hasForcible()) { 5246 hash = (37 * hash) + FORCIBLE_FIELD_NUMBER; 5247 hash = (53 * hash) + hashBoolean(getForcible()); 5248 } 5249 hash = (29 * hash) + getUnknownFields().hashCode(); 5250 memoizedHashCode = hash; 5251 return hash; 5252 } 5253 parseFrom( com.google.protobuf.ByteString data)5254 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5255 com.google.protobuf.ByteString data) 5256 throws com.google.protobuf.InvalidProtocolBufferException { 5257 return PARSER.parseFrom(data); 5258 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5259 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5260 com.google.protobuf.ByteString data, 5261 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5262 throws com.google.protobuf.InvalidProtocolBufferException { 5263 return PARSER.parseFrom(data, extensionRegistry); 5264 } parseFrom(byte[] data)5265 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(byte[] data) 5266 throws com.google.protobuf.InvalidProtocolBufferException { 5267 return PARSER.parseFrom(data); 5268 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5269 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5270 byte[] data, 5271 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5272 throws com.google.protobuf.InvalidProtocolBufferException { 5273 return PARSER.parseFrom(data, extensionRegistry); 5274 } parseFrom(java.io.InputStream input)5275 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom(java.io.InputStream input) 5276 throws java.io.IOException { 5277 return PARSER.parseFrom(input); 5278 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5279 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5280 java.io.InputStream input, 5281 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5282 throws java.io.IOException { 5283 return PARSER.parseFrom(input, extensionRegistry); 5284 } parseDelimitedFrom(java.io.InputStream input)5285 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom(java.io.InputStream input) 5286 throws java.io.IOException { 5287 return PARSER.parseDelimitedFrom(input); 5288 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5289 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseDelimitedFrom( 5290 java.io.InputStream input, 5291 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5292 throws java.io.IOException { 5293 return PARSER.parseDelimitedFrom(input, extensionRegistry); 5294 } parseFrom( com.google.protobuf.CodedInputStream input)5295 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5296 com.google.protobuf.CodedInputStream input) 5297 throws java.io.IOException { 5298 return PARSER.parseFrom(input); 5299 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5300 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parseFrom( 5301 com.google.protobuf.CodedInputStream input, 5302 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5303 throws java.io.IOException { 5304 return PARSER.parseFrom(input, extensionRegistry); 5305 } 5306 newBuilder()5307 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()5308 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest prototype)5309 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest prototype) { 5310 return newBuilder().mergeFrom(prototype); 5311 } toBuilder()5312 public Builder toBuilder() { return newBuilder(this); } 5313 5314 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)5315 protected Builder newBuilderForType( 5316 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5317 Builder builder = new Builder(parent); 5318 return builder; 5319 } 5320 /** 5321 * Protobuf type {@code DispatchMergingRegionsRequest} 5322 * 5323 * <pre> 5324 ** 5325 * Dispatch merging the specified regions. 5326 * </pre> 5327 */ 5328 public static final class Builder extends 5329 com.google.protobuf.GeneratedMessage.Builder<Builder> 5330 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequestOrBuilder { 5331 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()5332 getDescriptor() { 5333 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsRequest_descriptor; 5334 } 5335 5336 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()5337 internalGetFieldAccessorTable() { 5338 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsRequest_fieldAccessorTable 5339 .ensureFieldAccessorsInitialized( 5340 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.Builder.class); 5341 } 5342 5343 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.newBuilder() Builder()5344 private Builder() { 5345 maybeForceBuilderInitialization(); 5346 } 5347 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)5348 private Builder( 5349 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5350 super(parent); 5351 maybeForceBuilderInitialization(); 5352 } maybeForceBuilderInitialization()5353 private void maybeForceBuilderInitialization() { 5354 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 5355 getRegionAFieldBuilder(); 5356 getRegionBFieldBuilder(); 5357 } 5358 } create()5359 private static Builder create() { 5360 return new Builder(); 5361 } 5362 clear()5363 public Builder clear() { 5364 super.clear(); 5365 if (regionABuilder_ == null) { 5366 regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5367 } else { 5368 regionABuilder_.clear(); 5369 } 5370 bitField0_ = (bitField0_ & ~0x00000001); 5371 if (regionBBuilder_ == null) { 5372 regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5373 } else { 5374 regionBBuilder_.clear(); 5375 } 5376 bitField0_ = (bitField0_ & ~0x00000002); 5377 forcible_ = false; 5378 bitField0_ = (bitField0_ & ~0x00000004); 5379 return this; 5380 } 5381 clone()5382 public Builder clone() { 5383 return create().mergeFrom(buildPartial()); 5384 } 5385 5386 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()5387 getDescriptorForType() { 5388 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsRequest_descriptor; 5389 } 5390 getDefaultInstanceForType()5391 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest getDefaultInstanceForType() { 5392 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance(); 5393 } 5394 build()5395 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest build() { 5396 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = buildPartial(); 5397 if (!result.isInitialized()) { 5398 throw newUninitializedMessageException(result); 5399 } 5400 return result; 5401 } 5402 buildPartial()5403 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest buildPartial() { 5404 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest(this); 5405 int from_bitField0_ = bitField0_; 5406 int to_bitField0_ = 0; 5407 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 5408 to_bitField0_ |= 0x00000001; 5409 } 5410 if (regionABuilder_ == null) { 5411 result.regionA_ = regionA_; 5412 } else { 5413 result.regionA_ = regionABuilder_.build(); 5414 } 5415 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 5416 to_bitField0_ |= 0x00000002; 5417 } 5418 if (regionBBuilder_ == null) { 5419 result.regionB_ = regionB_; 5420 } else { 5421 result.regionB_ = regionBBuilder_.build(); 5422 } 5423 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 5424 to_bitField0_ |= 0x00000004; 5425 } 5426 result.forcible_ = forcible_; 5427 result.bitField0_ = to_bitField0_; 5428 onBuilt(); 5429 return result; 5430 } 5431 mergeFrom(com.google.protobuf.Message other)5432 public Builder mergeFrom(com.google.protobuf.Message other) { 5433 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) { 5434 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest)other); 5435 } else { 5436 super.mergeFrom(other); 5437 return this; 5438 } 5439 } 5440 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other)5441 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest other) { 5442 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest.getDefaultInstance()) return this; 5443 if (other.hasRegionA()) { 5444 mergeRegionA(other.getRegionA()); 5445 } 5446 if (other.hasRegionB()) { 5447 mergeRegionB(other.getRegionB()); 5448 } 5449 if (other.hasForcible()) { 5450 setForcible(other.getForcible()); 5451 } 5452 this.mergeUnknownFields(other.getUnknownFields()); 5453 return this; 5454 } 5455 isInitialized()5456 public final boolean isInitialized() { 5457 if (!hasRegionA()) { 5458 5459 return false; 5460 } 5461 if (!hasRegionB()) { 5462 5463 return false; 5464 } 5465 if (!getRegionA().isInitialized()) { 5466 5467 return false; 5468 } 5469 if (!getRegionB().isInitialized()) { 5470 5471 return false; 5472 } 5473 return true; 5474 } 5475 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5476 public Builder mergeFrom( 5477 com.google.protobuf.CodedInputStream input, 5478 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5479 throws java.io.IOException { 5480 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest parsedMessage = null; 5481 try { 5482 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 5483 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5484 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsRequest) e.getUnfinishedMessage(); 5485 throw e; 5486 } finally { 5487 if (parsedMessage != null) { 5488 mergeFrom(parsedMessage); 5489 } 5490 } 5491 return this; 5492 } 5493 private int bitField0_; 5494 5495 // required .RegionSpecifier region_a = 1; 5496 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5497 private com.google.protobuf.SingleFieldBuilder< 5498 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionABuilder_; 5499 /** 5500 * <code>required .RegionSpecifier region_a = 1;</code> 5501 */ hasRegionA()5502 public boolean hasRegionA() { 5503 return ((bitField0_ & 0x00000001) == 0x00000001); 5504 } 5505 /** 5506 * <code>required .RegionSpecifier region_a = 1;</code> 5507 */ getRegionA()5508 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionA() { 5509 if (regionABuilder_ == null) { 5510 return regionA_; 5511 } else { 5512 return regionABuilder_.getMessage(); 5513 } 5514 } 5515 /** 5516 * <code>required .RegionSpecifier region_a = 1;</code> 5517 */ setRegionA(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)5518 public Builder setRegionA(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 5519 if (regionABuilder_ == null) { 5520 if (value == null) { 5521 throw new NullPointerException(); 5522 } 5523 regionA_ = value; 5524 onChanged(); 5525 } else { 5526 regionABuilder_.setMessage(value); 5527 } 5528 bitField0_ |= 0x00000001; 5529 return this; 5530 } 5531 /** 5532 * <code>required .RegionSpecifier region_a = 1;</code> 5533 */ setRegionA( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)5534 public Builder setRegionA( 5535 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 5536 if (regionABuilder_ == null) { 5537 regionA_ = builderForValue.build(); 5538 onChanged(); 5539 } else { 5540 regionABuilder_.setMessage(builderForValue.build()); 5541 } 5542 bitField0_ |= 0x00000001; 5543 return this; 5544 } 5545 /** 5546 * <code>required .RegionSpecifier region_a = 1;</code> 5547 */ mergeRegionA(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)5548 public Builder mergeRegionA(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 5549 if (regionABuilder_ == null) { 5550 if (((bitField0_ & 0x00000001) == 0x00000001) && 5551 regionA_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 5552 regionA_ = 5553 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionA_).mergeFrom(value).buildPartial(); 5554 } else { 5555 regionA_ = value; 5556 } 5557 onChanged(); 5558 } else { 5559 regionABuilder_.mergeFrom(value); 5560 } 5561 bitField0_ |= 0x00000001; 5562 return this; 5563 } 5564 /** 5565 * <code>required .RegionSpecifier region_a = 1;</code> 5566 */ clearRegionA()5567 public Builder clearRegionA() { 5568 if (regionABuilder_ == null) { 5569 regionA_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5570 onChanged(); 5571 } else { 5572 regionABuilder_.clear(); 5573 } 5574 bitField0_ = (bitField0_ & ~0x00000001); 5575 return this; 5576 } 5577 /** 5578 * <code>required .RegionSpecifier region_a = 1;</code> 5579 */ getRegionABuilder()5580 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionABuilder() { 5581 bitField0_ |= 0x00000001; 5582 onChanged(); 5583 return getRegionAFieldBuilder().getBuilder(); 5584 } 5585 /** 5586 * <code>required .RegionSpecifier region_a = 1;</code> 5587 */ getRegionAOrBuilder()5588 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionAOrBuilder() { 5589 if (regionABuilder_ != null) { 5590 return regionABuilder_.getMessageOrBuilder(); 5591 } else { 5592 return regionA_; 5593 } 5594 } 5595 /** 5596 * <code>required .RegionSpecifier region_a = 1;</code> 5597 */ 5598 private com.google.protobuf.SingleFieldBuilder< 5599 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionAFieldBuilder()5600 getRegionAFieldBuilder() { 5601 if (regionABuilder_ == null) { 5602 regionABuilder_ = new com.google.protobuf.SingleFieldBuilder< 5603 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 5604 regionA_, 5605 getParentForChildren(), 5606 isClean()); 5607 regionA_ = null; 5608 } 5609 return regionABuilder_; 5610 } 5611 5612 // required .RegionSpecifier region_b = 2; 5613 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5614 private com.google.protobuf.SingleFieldBuilder< 5615 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBBuilder_; 5616 /** 5617 * <code>required .RegionSpecifier region_b = 2;</code> 5618 */ hasRegionB()5619 public boolean hasRegionB() { 5620 return ((bitField0_ & 0x00000002) == 0x00000002); 5621 } 5622 /** 5623 * <code>required .RegionSpecifier region_b = 2;</code> 5624 */ getRegionB()5625 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegionB() { 5626 if (regionBBuilder_ == null) { 5627 return regionB_; 5628 } else { 5629 return regionBBuilder_.getMessage(); 5630 } 5631 } 5632 /** 5633 * <code>required .RegionSpecifier region_b = 2;</code> 5634 */ setRegionB(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)5635 public Builder setRegionB(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 5636 if (regionBBuilder_ == null) { 5637 if (value == null) { 5638 throw new NullPointerException(); 5639 } 5640 regionB_ = value; 5641 onChanged(); 5642 } else { 5643 regionBBuilder_.setMessage(value); 5644 } 5645 bitField0_ |= 0x00000002; 5646 return this; 5647 } 5648 /** 5649 * <code>required .RegionSpecifier region_b = 2;</code> 5650 */ setRegionB( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)5651 public Builder setRegionB( 5652 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 5653 if (regionBBuilder_ == null) { 5654 regionB_ = builderForValue.build(); 5655 onChanged(); 5656 } else { 5657 regionBBuilder_.setMessage(builderForValue.build()); 5658 } 5659 bitField0_ |= 0x00000002; 5660 return this; 5661 } 5662 /** 5663 * <code>required .RegionSpecifier region_b = 2;</code> 5664 */ mergeRegionB(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)5665 public Builder mergeRegionB(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 5666 if (regionBBuilder_ == null) { 5667 if (((bitField0_ & 0x00000002) == 0x00000002) && 5668 regionB_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 5669 regionB_ = 5670 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionB_).mergeFrom(value).buildPartial(); 5671 } else { 5672 regionB_ = value; 5673 } 5674 onChanged(); 5675 } else { 5676 regionBBuilder_.mergeFrom(value); 5677 } 5678 bitField0_ |= 0x00000002; 5679 return this; 5680 } 5681 /** 5682 * <code>required .RegionSpecifier region_b = 2;</code> 5683 */ clearRegionB()5684 public Builder clearRegionB() { 5685 if (regionBBuilder_ == null) { 5686 regionB_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 5687 onChanged(); 5688 } else { 5689 regionBBuilder_.clear(); 5690 } 5691 bitField0_ = (bitField0_ & ~0x00000002); 5692 return this; 5693 } 5694 /** 5695 * <code>required .RegionSpecifier region_b = 2;</code> 5696 */ getRegionBBuilder()5697 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBBuilder() { 5698 bitField0_ |= 0x00000002; 5699 onChanged(); 5700 return getRegionBFieldBuilder().getBuilder(); 5701 } 5702 /** 5703 * <code>required .RegionSpecifier region_b = 2;</code> 5704 */ getRegionBOrBuilder()5705 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionBOrBuilder() { 5706 if (regionBBuilder_ != null) { 5707 return regionBBuilder_.getMessageOrBuilder(); 5708 } else { 5709 return regionB_; 5710 } 5711 } 5712 /** 5713 * <code>required .RegionSpecifier region_b = 2;</code> 5714 */ 5715 private com.google.protobuf.SingleFieldBuilder< 5716 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionBFieldBuilder()5717 getRegionBFieldBuilder() { 5718 if (regionBBuilder_ == null) { 5719 regionBBuilder_ = new com.google.protobuf.SingleFieldBuilder< 5720 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 5721 regionB_, 5722 getParentForChildren(), 5723 isClean()); 5724 regionB_ = null; 5725 } 5726 return regionBBuilder_; 5727 } 5728 5729 // optional bool forcible = 3 [default = false]; 5730 private boolean forcible_ ; 5731 /** 5732 * <code>optional bool forcible = 3 [default = false];</code> 5733 */ hasForcible()5734 public boolean hasForcible() { 5735 return ((bitField0_ & 0x00000004) == 0x00000004); 5736 } 5737 /** 5738 * <code>optional bool forcible = 3 [default = false];</code> 5739 */ getForcible()5740 public boolean getForcible() { 5741 return forcible_; 5742 } 5743 /** 5744 * <code>optional bool forcible = 3 [default = false];</code> 5745 */ setForcible(boolean value)5746 public Builder setForcible(boolean value) { 5747 bitField0_ |= 0x00000004; 5748 forcible_ = value; 5749 onChanged(); 5750 return this; 5751 } 5752 /** 5753 * <code>optional bool forcible = 3 [default = false];</code> 5754 */ clearForcible()5755 public Builder clearForcible() { 5756 bitField0_ = (bitField0_ & ~0x00000004); 5757 forcible_ = false; 5758 onChanged(); 5759 return this; 5760 } 5761 5762 // @@protoc_insertion_point(builder_scope:DispatchMergingRegionsRequest) 5763 } 5764 5765 static { 5766 defaultInstance = new DispatchMergingRegionsRequest(true); defaultInstance.initFields()5767 defaultInstance.initFields(); 5768 } 5769 5770 // @@protoc_insertion_point(class_scope:DispatchMergingRegionsRequest) 5771 } 5772 5773 public interface DispatchMergingRegionsResponseOrBuilder 5774 extends com.google.protobuf.MessageOrBuilder { 5775 } 5776 /** 5777 * Protobuf type {@code DispatchMergingRegionsResponse} 5778 */ 5779 public static final class DispatchMergingRegionsResponse extends 5780 com.google.protobuf.GeneratedMessage 5781 implements DispatchMergingRegionsResponseOrBuilder { 5782 // Use DispatchMergingRegionsResponse.newBuilder() to construct. DispatchMergingRegionsResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)5783 private DispatchMergingRegionsResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 5784 super(builder); 5785 this.unknownFields = builder.getUnknownFields(); 5786 } DispatchMergingRegionsResponse(boolean noInit)5787 private DispatchMergingRegionsResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 5788 5789 private static final DispatchMergingRegionsResponse defaultInstance; getDefaultInstance()5790 public static DispatchMergingRegionsResponse getDefaultInstance() { 5791 return defaultInstance; 5792 } 5793 getDefaultInstanceForType()5794 public DispatchMergingRegionsResponse getDefaultInstanceForType() { 5795 return defaultInstance; 5796 } 5797 5798 private final com.google.protobuf.UnknownFieldSet unknownFields; 5799 @java.lang.Override 5800 public final com.google.protobuf.UnknownFieldSet getUnknownFields()5801 getUnknownFields() { 5802 return this.unknownFields; 5803 } DispatchMergingRegionsResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5804 private DispatchMergingRegionsResponse( 5805 com.google.protobuf.CodedInputStream input, 5806 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5807 throws com.google.protobuf.InvalidProtocolBufferException { 5808 initFields(); 5809 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 5810 com.google.protobuf.UnknownFieldSet.newBuilder(); 5811 try { 5812 boolean done = false; 5813 while (!done) { 5814 int tag = input.readTag(); 5815 switch (tag) { 5816 case 0: 5817 done = true; 5818 break; 5819 default: { 5820 if (!parseUnknownField(input, unknownFields, 5821 extensionRegistry, tag)) { 5822 done = true; 5823 } 5824 break; 5825 } 5826 } 5827 } 5828 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 5829 throw e.setUnfinishedMessage(this); 5830 } catch (java.io.IOException e) { 5831 throw new com.google.protobuf.InvalidProtocolBufferException( 5832 e.getMessage()).setUnfinishedMessage(this); 5833 } finally { 5834 this.unknownFields = unknownFields.build(); 5835 makeExtensionsImmutable(); 5836 } 5837 } 5838 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()5839 getDescriptor() { 5840 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsResponse_descriptor; 5841 } 5842 5843 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()5844 internalGetFieldAccessorTable() { 5845 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsResponse_fieldAccessorTable 5846 .ensureFieldAccessorsInitialized( 5847 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class); 5848 } 5849 5850 public static com.google.protobuf.Parser<DispatchMergingRegionsResponse> PARSER = 5851 new com.google.protobuf.AbstractParser<DispatchMergingRegionsResponse>() { 5852 public DispatchMergingRegionsResponse parsePartialFrom( 5853 com.google.protobuf.CodedInputStream input, 5854 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5855 throws com.google.protobuf.InvalidProtocolBufferException { 5856 return new DispatchMergingRegionsResponse(input, extensionRegistry); 5857 } 5858 }; 5859 5860 @java.lang.Override getParserForType()5861 public com.google.protobuf.Parser<DispatchMergingRegionsResponse> getParserForType() { 5862 return PARSER; 5863 } 5864 initFields()5865 private void initFields() { 5866 } 5867 private byte memoizedIsInitialized = -1; isInitialized()5868 public final boolean isInitialized() { 5869 byte isInitialized = memoizedIsInitialized; 5870 if (isInitialized != -1) return isInitialized == 1; 5871 5872 memoizedIsInitialized = 1; 5873 return true; 5874 } 5875 writeTo(com.google.protobuf.CodedOutputStream output)5876 public void writeTo(com.google.protobuf.CodedOutputStream output) 5877 throws java.io.IOException { 5878 getSerializedSize(); 5879 getUnknownFields().writeTo(output); 5880 } 5881 5882 private int memoizedSerializedSize = -1; getSerializedSize()5883 public int getSerializedSize() { 5884 int size = memoizedSerializedSize; 5885 if (size != -1) return size; 5886 5887 size = 0; 5888 size += getUnknownFields().getSerializedSize(); 5889 memoizedSerializedSize = size; 5890 return size; 5891 } 5892 5893 private static final long serialVersionUID = 0L; 5894 @java.lang.Override writeReplace()5895 protected java.lang.Object writeReplace() 5896 throws java.io.ObjectStreamException { 5897 return super.writeReplace(); 5898 } 5899 5900 @java.lang.Override equals(final java.lang.Object obj)5901 public boolean equals(final java.lang.Object obj) { 5902 if (obj == this) { 5903 return true; 5904 } 5905 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)) { 5906 return super.equals(obj); 5907 } 5908 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) obj; 5909 5910 boolean result = true; 5911 result = result && 5912 getUnknownFields().equals(other.getUnknownFields()); 5913 return result; 5914 } 5915 5916 private int memoizedHashCode = 0; 5917 @java.lang.Override hashCode()5918 public int hashCode() { 5919 if (memoizedHashCode != 0) { 5920 return memoizedHashCode; 5921 } 5922 int hash = 41; 5923 hash = (19 * hash) + getDescriptorForType().hashCode(); 5924 hash = (29 * hash) + getUnknownFields().hashCode(); 5925 memoizedHashCode = hash; 5926 return hash; 5927 } 5928 parseFrom( com.google.protobuf.ByteString data)5929 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5930 com.google.protobuf.ByteString data) 5931 throws com.google.protobuf.InvalidProtocolBufferException { 5932 return PARSER.parseFrom(data); 5933 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5934 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5935 com.google.protobuf.ByteString data, 5936 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5937 throws com.google.protobuf.InvalidProtocolBufferException { 5938 return PARSER.parseFrom(data, extensionRegistry); 5939 } parseFrom(byte[] data)5940 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(byte[] data) 5941 throws com.google.protobuf.InvalidProtocolBufferException { 5942 return PARSER.parseFrom(data); 5943 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5944 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5945 byte[] data, 5946 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5947 throws com.google.protobuf.InvalidProtocolBufferException { 5948 return PARSER.parseFrom(data, extensionRegistry); 5949 } parseFrom(java.io.InputStream input)5950 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom(java.io.InputStream input) 5951 throws java.io.IOException { 5952 return PARSER.parseFrom(input); 5953 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5954 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5955 java.io.InputStream input, 5956 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5957 throws java.io.IOException { 5958 return PARSER.parseFrom(input, extensionRegistry); 5959 } parseDelimitedFrom(java.io.InputStream input)5960 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom(java.io.InputStream input) 5961 throws java.io.IOException { 5962 return PARSER.parseDelimitedFrom(input); 5963 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5964 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseDelimitedFrom( 5965 java.io.InputStream input, 5966 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5967 throws java.io.IOException { 5968 return PARSER.parseDelimitedFrom(input, extensionRegistry); 5969 } parseFrom( com.google.protobuf.CodedInputStream input)5970 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5971 com.google.protobuf.CodedInputStream input) 5972 throws java.io.IOException { 5973 return PARSER.parseFrom(input); 5974 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)5975 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parseFrom( 5976 com.google.protobuf.CodedInputStream input, 5977 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 5978 throws java.io.IOException { 5979 return PARSER.parseFrom(input, extensionRegistry); 5980 } 5981 newBuilder()5982 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()5983 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse prototype)5984 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse prototype) { 5985 return newBuilder().mergeFrom(prototype); 5986 } toBuilder()5987 public Builder toBuilder() { return newBuilder(this); } 5988 5989 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)5990 protected Builder newBuilderForType( 5991 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 5992 Builder builder = new Builder(parent); 5993 return builder; 5994 } 5995 /** 5996 * Protobuf type {@code DispatchMergingRegionsResponse} 5997 */ 5998 public static final class Builder extends 5999 com.google.protobuf.GeneratedMessage.Builder<Builder> 6000 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponseOrBuilder { 6001 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()6002 getDescriptor() { 6003 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsResponse_descriptor; 6004 } 6005 6006 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()6007 internalGetFieldAccessorTable() { 6008 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsResponse_fieldAccessorTable 6009 .ensureFieldAccessorsInitialized( 6010 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.Builder.class); 6011 } 6012 6013 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.newBuilder() Builder()6014 private Builder() { 6015 maybeForceBuilderInitialization(); 6016 } 6017 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)6018 private Builder( 6019 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6020 super(parent); 6021 maybeForceBuilderInitialization(); 6022 } maybeForceBuilderInitialization()6023 private void maybeForceBuilderInitialization() { 6024 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6025 } 6026 } create()6027 private static Builder create() { 6028 return new Builder(); 6029 } 6030 clear()6031 public Builder clear() { 6032 super.clear(); 6033 return this; 6034 } 6035 clone()6036 public Builder clone() { 6037 return create().mergeFrom(buildPartial()); 6038 } 6039 6040 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()6041 getDescriptorForType() { 6042 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_DispatchMergingRegionsResponse_descriptor; 6043 } 6044 getDefaultInstanceForType()6045 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse getDefaultInstanceForType() { 6046 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance(); 6047 } 6048 build()6049 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse build() { 6050 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = buildPartial(); 6051 if (!result.isInitialized()) { 6052 throw newUninitializedMessageException(result); 6053 } 6054 return result; 6055 } 6056 buildPartial()6057 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse buildPartial() { 6058 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse(this); 6059 onBuilt(); 6060 return result; 6061 } 6062 mergeFrom(com.google.protobuf.Message other)6063 public Builder mergeFrom(com.google.protobuf.Message other) { 6064 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) { 6065 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse)other); 6066 } else { 6067 super.mergeFrom(other); 6068 return this; 6069 } 6070 } 6071 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other)6072 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse other) { 6073 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse.getDefaultInstance()) return this; 6074 this.mergeUnknownFields(other.getUnknownFields()); 6075 return this; 6076 } 6077 isInitialized()6078 public final boolean isInitialized() { 6079 return true; 6080 } 6081 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6082 public Builder mergeFrom( 6083 com.google.protobuf.CodedInputStream input, 6084 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6085 throws java.io.IOException { 6086 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse parsedMessage = null; 6087 try { 6088 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6089 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6090 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse) e.getUnfinishedMessage(); 6091 throw e; 6092 } finally { 6093 if (parsedMessage != null) { 6094 mergeFrom(parsedMessage); 6095 } 6096 } 6097 return this; 6098 } 6099 6100 // @@protoc_insertion_point(builder_scope:DispatchMergingRegionsResponse) 6101 } 6102 6103 static { 6104 defaultInstance = new DispatchMergingRegionsResponse(true); defaultInstance.initFields()6105 defaultInstance.initFields(); 6106 } 6107 6108 // @@protoc_insertion_point(class_scope:DispatchMergingRegionsResponse) 6109 } 6110 6111 public interface AssignRegionRequestOrBuilder 6112 extends com.google.protobuf.MessageOrBuilder { 6113 6114 // required .RegionSpecifier region = 1; 6115 /** 6116 * <code>required .RegionSpecifier region = 1;</code> 6117 */ hasRegion()6118 boolean hasRegion(); 6119 /** 6120 * <code>required .RegionSpecifier region = 1;</code> 6121 */ getRegion()6122 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); 6123 /** 6124 * <code>required .RegionSpecifier region = 1;</code> 6125 */ getRegionOrBuilder()6126 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); 6127 } 6128 /** 6129 * Protobuf type {@code AssignRegionRequest} 6130 */ 6131 public static final class AssignRegionRequest extends 6132 com.google.protobuf.GeneratedMessage 6133 implements AssignRegionRequestOrBuilder { 6134 // Use AssignRegionRequest.newBuilder() to construct. AssignRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)6135 private AssignRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 6136 super(builder); 6137 this.unknownFields = builder.getUnknownFields(); 6138 } AssignRegionRequest(boolean noInit)6139 private AssignRegionRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 6140 6141 private static final AssignRegionRequest defaultInstance; getDefaultInstance()6142 public static AssignRegionRequest getDefaultInstance() { 6143 return defaultInstance; 6144 } 6145 getDefaultInstanceForType()6146 public AssignRegionRequest getDefaultInstanceForType() { 6147 return defaultInstance; 6148 } 6149 6150 private final com.google.protobuf.UnknownFieldSet unknownFields; 6151 @java.lang.Override 6152 public final com.google.protobuf.UnknownFieldSet getUnknownFields()6153 getUnknownFields() { 6154 return this.unknownFields; 6155 } AssignRegionRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6156 private AssignRegionRequest( 6157 com.google.protobuf.CodedInputStream input, 6158 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6159 throws com.google.protobuf.InvalidProtocolBufferException { 6160 initFields(); 6161 int mutable_bitField0_ = 0; 6162 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 6163 com.google.protobuf.UnknownFieldSet.newBuilder(); 6164 try { 6165 boolean done = false; 6166 while (!done) { 6167 int tag = input.readTag(); 6168 switch (tag) { 6169 case 0: 6170 done = true; 6171 break; 6172 default: { 6173 if (!parseUnknownField(input, unknownFields, 6174 extensionRegistry, tag)) { 6175 done = true; 6176 } 6177 break; 6178 } 6179 case 10: { 6180 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 6181 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6182 subBuilder = region_.toBuilder(); 6183 } 6184 region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 6185 if (subBuilder != null) { 6186 subBuilder.mergeFrom(region_); 6187 region_ = subBuilder.buildPartial(); 6188 } 6189 bitField0_ |= 0x00000001; 6190 break; 6191 } 6192 } 6193 } 6194 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6195 throw e.setUnfinishedMessage(this); 6196 } catch (java.io.IOException e) { 6197 throw new com.google.protobuf.InvalidProtocolBufferException( 6198 e.getMessage()).setUnfinishedMessage(this); 6199 } finally { 6200 this.unknownFields = unknownFields.build(); 6201 makeExtensionsImmutable(); 6202 } 6203 } 6204 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()6205 getDescriptor() { 6206 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionRequest_descriptor; 6207 } 6208 6209 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()6210 internalGetFieldAccessorTable() { 6211 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionRequest_fieldAccessorTable 6212 .ensureFieldAccessorsInitialized( 6213 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.Builder.class); 6214 } 6215 6216 public static com.google.protobuf.Parser<AssignRegionRequest> PARSER = 6217 new com.google.protobuf.AbstractParser<AssignRegionRequest>() { 6218 public AssignRegionRequest parsePartialFrom( 6219 com.google.protobuf.CodedInputStream input, 6220 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6221 throws com.google.protobuf.InvalidProtocolBufferException { 6222 return new AssignRegionRequest(input, extensionRegistry); 6223 } 6224 }; 6225 6226 @java.lang.Override getParserForType()6227 public com.google.protobuf.Parser<AssignRegionRequest> getParserForType() { 6228 return PARSER; 6229 } 6230 6231 private int bitField0_; 6232 // required .RegionSpecifier region = 1; 6233 public static final int REGION_FIELD_NUMBER = 1; 6234 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_; 6235 /** 6236 * <code>required .RegionSpecifier region = 1;</code> 6237 */ hasRegion()6238 public boolean hasRegion() { 6239 return ((bitField0_ & 0x00000001) == 0x00000001); 6240 } 6241 /** 6242 * <code>required .RegionSpecifier region = 1;</code> 6243 */ getRegion()6244 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 6245 return region_; 6246 } 6247 /** 6248 * <code>required .RegionSpecifier region = 1;</code> 6249 */ getRegionOrBuilder()6250 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 6251 return region_; 6252 } 6253 initFields()6254 private void initFields() { 6255 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 6256 } 6257 private byte memoizedIsInitialized = -1; isInitialized()6258 public final boolean isInitialized() { 6259 byte isInitialized = memoizedIsInitialized; 6260 if (isInitialized != -1) return isInitialized == 1; 6261 6262 if (!hasRegion()) { 6263 memoizedIsInitialized = 0; 6264 return false; 6265 } 6266 if (!getRegion().isInitialized()) { 6267 memoizedIsInitialized = 0; 6268 return false; 6269 } 6270 memoizedIsInitialized = 1; 6271 return true; 6272 } 6273 writeTo(com.google.protobuf.CodedOutputStream output)6274 public void writeTo(com.google.protobuf.CodedOutputStream output) 6275 throws java.io.IOException { 6276 getSerializedSize(); 6277 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6278 output.writeMessage(1, region_); 6279 } 6280 getUnknownFields().writeTo(output); 6281 } 6282 6283 private int memoizedSerializedSize = -1; getSerializedSize()6284 public int getSerializedSize() { 6285 int size = memoizedSerializedSize; 6286 if (size != -1) return size; 6287 6288 size = 0; 6289 if (((bitField0_ & 0x00000001) == 0x00000001)) { 6290 size += com.google.protobuf.CodedOutputStream 6291 .computeMessageSize(1, region_); 6292 } 6293 size += getUnknownFields().getSerializedSize(); 6294 memoizedSerializedSize = size; 6295 return size; 6296 } 6297 6298 private static final long serialVersionUID = 0L; 6299 @java.lang.Override writeReplace()6300 protected java.lang.Object writeReplace() 6301 throws java.io.ObjectStreamException { 6302 return super.writeReplace(); 6303 } 6304 6305 @java.lang.Override equals(final java.lang.Object obj)6306 public boolean equals(final java.lang.Object obj) { 6307 if (obj == this) { 6308 return true; 6309 } 6310 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest)) { 6311 return super.equals(obj); 6312 } 6313 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest) obj; 6314 6315 boolean result = true; 6316 result = result && (hasRegion() == other.hasRegion()); 6317 if (hasRegion()) { 6318 result = result && getRegion() 6319 .equals(other.getRegion()); 6320 } 6321 result = result && 6322 getUnknownFields().equals(other.getUnknownFields()); 6323 return result; 6324 } 6325 6326 private int memoizedHashCode = 0; 6327 @java.lang.Override hashCode()6328 public int hashCode() { 6329 if (memoizedHashCode != 0) { 6330 return memoizedHashCode; 6331 } 6332 int hash = 41; 6333 hash = (19 * hash) + getDescriptorForType().hashCode(); 6334 if (hasRegion()) { 6335 hash = (37 * hash) + REGION_FIELD_NUMBER; 6336 hash = (53 * hash) + getRegion().hashCode(); 6337 } 6338 hash = (29 * hash) + getUnknownFields().hashCode(); 6339 memoizedHashCode = hash; 6340 return hash; 6341 } 6342 parseFrom( com.google.protobuf.ByteString data)6343 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6344 com.google.protobuf.ByteString data) 6345 throws com.google.protobuf.InvalidProtocolBufferException { 6346 return PARSER.parseFrom(data); 6347 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6348 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6349 com.google.protobuf.ByteString data, 6350 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6351 throws com.google.protobuf.InvalidProtocolBufferException { 6352 return PARSER.parseFrom(data, extensionRegistry); 6353 } parseFrom(byte[] data)6354 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom(byte[] data) 6355 throws com.google.protobuf.InvalidProtocolBufferException { 6356 return PARSER.parseFrom(data); 6357 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6358 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6359 byte[] data, 6360 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6361 throws com.google.protobuf.InvalidProtocolBufferException { 6362 return PARSER.parseFrom(data, extensionRegistry); 6363 } parseFrom(java.io.InputStream input)6364 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom(java.io.InputStream input) 6365 throws java.io.IOException { 6366 return PARSER.parseFrom(input); 6367 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6368 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6369 java.io.InputStream input, 6370 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6371 throws java.io.IOException { 6372 return PARSER.parseFrom(input, extensionRegistry); 6373 } parseDelimitedFrom(java.io.InputStream input)6374 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseDelimitedFrom(java.io.InputStream input) 6375 throws java.io.IOException { 6376 return PARSER.parseDelimitedFrom(input); 6377 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6378 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseDelimitedFrom( 6379 java.io.InputStream input, 6380 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6381 throws java.io.IOException { 6382 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6383 } parseFrom( com.google.protobuf.CodedInputStream input)6384 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6385 com.google.protobuf.CodedInputStream input) 6386 throws java.io.IOException { 6387 return PARSER.parseFrom(input); 6388 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6389 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( 6390 com.google.protobuf.CodedInputStream input, 6391 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6392 throws java.io.IOException { 6393 return PARSER.parseFrom(input, extensionRegistry); 6394 } 6395 newBuilder()6396 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()6397 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest prototype)6398 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest prototype) { 6399 return newBuilder().mergeFrom(prototype); 6400 } toBuilder()6401 public Builder toBuilder() { return newBuilder(this); } 6402 6403 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)6404 protected Builder newBuilderForType( 6405 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6406 Builder builder = new Builder(parent); 6407 return builder; 6408 } 6409 /** 6410 * Protobuf type {@code AssignRegionRequest} 6411 */ 6412 public static final class Builder extends 6413 com.google.protobuf.GeneratedMessage.Builder<Builder> 6414 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequestOrBuilder { 6415 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()6416 getDescriptor() { 6417 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionRequest_descriptor; 6418 } 6419 6420 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()6421 internalGetFieldAccessorTable() { 6422 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionRequest_fieldAccessorTable 6423 .ensureFieldAccessorsInitialized( 6424 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.Builder.class); 6425 } 6426 6427 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.newBuilder() Builder()6428 private Builder() { 6429 maybeForceBuilderInitialization(); 6430 } 6431 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)6432 private Builder( 6433 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6434 super(parent); 6435 maybeForceBuilderInitialization(); 6436 } maybeForceBuilderInitialization()6437 private void maybeForceBuilderInitialization() { 6438 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6439 getRegionFieldBuilder(); 6440 } 6441 } create()6442 private static Builder create() { 6443 return new Builder(); 6444 } 6445 clear()6446 public Builder clear() { 6447 super.clear(); 6448 if (regionBuilder_ == null) { 6449 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 6450 } else { 6451 regionBuilder_.clear(); 6452 } 6453 bitField0_ = (bitField0_ & ~0x00000001); 6454 return this; 6455 } 6456 clone()6457 public Builder clone() { 6458 return create().mergeFrom(buildPartial()); 6459 } 6460 6461 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()6462 getDescriptorForType() { 6463 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionRequest_descriptor; 6464 } 6465 getDefaultInstanceForType()6466 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest getDefaultInstanceForType() { 6467 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); 6468 } 6469 build()6470 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest build() { 6471 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest result = buildPartial(); 6472 if (!result.isInitialized()) { 6473 throw newUninitializedMessageException(result); 6474 } 6475 return result; 6476 } 6477 buildPartial()6478 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest buildPartial() { 6479 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest(this); 6480 int from_bitField0_ = bitField0_; 6481 int to_bitField0_ = 0; 6482 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 6483 to_bitField0_ |= 0x00000001; 6484 } 6485 if (regionBuilder_ == null) { 6486 result.region_ = region_; 6487 } else { 6488 result.region_ = regionBuilder_.build(); 6489 } 6490 result.bitField0_ = to_bitField0_; 6491 onBuilt(); 6492 return result; 6493 } 6494 mergeFrom(com.google.protobuf.Message other)6495 public Builder mergeFrom(com.google.protobuf.Message other) { 6496 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest) { 6497 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest)other); 6498 } else { 6499 super.mergeFrom(other); 6500 return this; 6501 } 6502 } 6503 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest other)6504 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest other) { 6505 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance()) return this; 6506 if (other.hasRegion()) { 6507 mergeRegion(other.getRegion()); 6508 } 6509 this.mergeUnknownFields(other.getUnknownFields()); 6510 return this; 6511 } 6512 isInitialized()6513 public final boolean isInitialized() { 6514 if (!hasRegion()) { 6515 6516 return false; 6517 } 6518 if (!getRegion().isInitialized()) { 6519 6520 return false; 6521 } 6522 return true; 6523 } 6524 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6525 public Builder mergeFrom( 6526 com.google.protobuf.CodedInputStream input, 6527 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6528 throws java.io.IOException { 6529 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest parsedMessage = null; 6530 try { 6531 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6532 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6533 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionRequest) e.getUnfinishedMessage(); 6534 throw e; 6535 } finally { 6536 if (parsedMessage != null) { 6537 mergeFrom(parsedMessage); 6538 } 6539 } 6540 return this; 6541 } 6542 private int bitField0_; 6543 6544 // required .RegionSpecifier region = 1; 6545 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 6546 private com.google.protobuf.SingleFieldBuilder< 6547 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; 6548 /** 6549 * <code>required .RegionSpecifier region = 1;</code> 6550 */ hasRegion()6551 public boolean hasRegion() { 6552 return ((bitField0_ & 0x00000001) == 0x00000001); 6553 } 6554 /** 6555 * <code>required .RegionSpecifier region = 1;</code> 6556 */ getRegion()6557 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 6558 if (regionBuilder_ == null) { 6559 return region_; 6560 } else { 6561 return regionBuilder_.getMessage(); 6562 } 6563 } 6564 /** 6565 * <code>required .RegionSpecifier region = 1;</code> 6566 */ setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)6567 public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 6568 if (regionBuilder_ == null) { 6569 if (value == null) { 6570 throw new NullPointerException(); 6571 } 6572 region_ = value; 6573 onChanged(); 6574 } else { 6575 regionBuilder_.setMessage(value); 6576 } 6577 bitField0_ |= 0x00000001; 6578 return this; 6579 } 6580 /** 6581 * <code>required .RegionSpecifier region = 1;</code> 6582 */ setRegion( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)6583 public Builder setRegion( 6584 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 6585 if (regionBuilder_ == null) { 6586 region_ = builderForValue.build(); 6587 onChanged(); 6588 } else { 6589 regionBuilder_.setMessage(builderForValue.build()); 6590 } 6591 bitField0_ |= 0x00000001; 6592 return this; 6593 } 6594 /** 6595 * <code>required .RegionSpecifier region = 1;</code> 6596 */ mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)6597 public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 6598 if (regionBuilder_ == null) { 6599 if (((bitField0_ & 0x00000001) == 0x00000001) && 6600 region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 6601 region_ = 6602 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); 6603 } else { 6604 region_ = value; 6605 } 6606 onChanged(); 6607 } else { 6608 regionBuilder_.mergeFrom(value); 6609 } 6610 bitField0_ |= 0x00000001; 6611 return this; 6612 } 6613 /** 6614 * <code>required .RegionSpecifier region = 1;</code> 6615 */ clearRegion()6616 public Builder clearRegion() { 6617 if (regionBuilder_ == null) { 6618 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 6619 onChanged(); 6620 } else { 6621 regionBuilder_.clear(); 6622 } 6623 bitField0_ = (bitField0_ & ~0x00000001); 6624 return this; 6625 } 6626 /** 6627 * <code>required .RegionSpecifier region = 1;</code> 6628 */ getRegionBuilder()6629 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { 6630 bitField0_ |= 0x00000001; 6631 onChanged(); 6632 return getRegionFieldBuilder().getBuilder(); 6633 } 6634 /** 6635 * <code>required .RegionSpecifier region = 1;</code> 6636 */ getRegionOrBuilder()6637 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 6638 if (regionBuilder_ != null) { 6639 return regionBuilder_.getMessageOrBuilder(); 6640 } else { 6641 return region_; 6642 } 6643 } 6644 /** 6645 * <code>required .RegionSpecifier region = 1;</code> 6646 */ 6647 private com.google.protobuf.SingleFieldBuilder< 6648 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder()6649 getRegionFieldBuilder() { 6650 if (regionBuilder_ == null) { 6651 regionBuilder_ = new com.google.protobuf.SingleFieldBuilder< 6652 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 6653 region_, 6654 getParentForChildren(), 6655 isClean()); 6656 region_ = null; 6657 } 6658 return regionBuilder_; 6659 } 6660 6661 // @@protoc_insertion_point(builder_scope:AssignRegionRequest) 6662 } 6663 6664 static { 6665 defaultInstance = new AssignRegionRequest(true); defaultInstance.initFields()6666 defaultInstance.initFields(); 6667 } 6668 6669 // @@protoc_insertion_point(class_scope:AssignRegionRequest) 6670 } 6671 6672 public interface AssignRegionResponseOrBuilder 6673 extends com.google.protobuf.MessageOrBuilder { 6674 } 6675 /** 6676 * Protobuf type {@code AssignRegionResponse} 6677 */ 6678 public static final class AssignRegionResponse extends 6679 com.google.protobuf.GeneratedMessage 6680 implements AssignRegionResponseOrBuilder { 6681 // Use AssignRegionResponse.newBuilder() to construct. AssignRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)6682 private AssignRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 6683 super(builder); 6684 this.unknownFields = builder.getUnknownFields(); 6685 } AssignRegionResponse(boolean noInit)6686 private AssignRegionResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 6687 6688 private static final AssignRegionResponse defaultInstance; getDefaultInstance()6689 public static AssignRegionResponse getDefaultInstance() { 6690 return defaultInstance; 6691 } 6692 getDefaultInstanceForType()6693 public AssignRegionResponse getDefaultInstanceForType() { 6694 return defaultInstance; 6695 } 6696 6697 private final com.google.protobuf.UnknownFieldSet unknownFields; 6698 @java.lang.Override 6699 public final com.google.protobuf.UnknownFieldSet getUnknownFields()6700 getUnknownFields() { 6701 return this.unknownFields; 6702 } AssignRegionResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6703 private AssignRegionResponse( 6704 com.google.protobuf.CodedInputStream input, 6705 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6706 throws com.google.protobuf.InvalidProtocolBufferException { 6707 initFields(); 6708 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 6709 com.google.protobuf.UnknownFieldSet.newBuilder(); 6710 try { 6711 boolean done = false; 6712 while (!done) { 6713 int tag = input.readTag(); 6714 switch (tag) { 6715 case 0: 6716 done = true; 6717 break; 6718 default: { 6719 if (!parseUnknownField(input, unknownFields, 6720 extensionRegistry, tag)) { 6721 done = true; 6722 } 6723 break; 6724 } 6725 } 6726 } 6727 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6728 throw e.setUnfinishedMessage(this); 6729 } catch (java.io.IOException e) { 6730 throw new com.google.protobuf.InvalidProtocolBufferException( 6731 e.getMessage()).setUnfinishedMessage(this); 6732 } finally { 6733 this.unknownFields = unknownFields.build(); 6734 makeExtensionsImmutable(); 6735 } 6736 } 6737 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()6738 getDescriptor() { 6739 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionResponse_descriptor; 6740 } 6741 6742 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()6743 internalGetFieldAccessorTable() { 6744 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionResponse_fieldAccessorTable 6745 .ensureFieldAccessorsInitialized( 6746 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.Builder.class); 6747 } 6748 6749 public static com.google.protobuf.Parser<AssignRegionResponse> PARSER = 6750 new com.google.protobuf.AbstractParser<AssignRegionResponse>() { 6751 public AssignRegionResponse parsePartialFrom( 6752 com.google.protobuf.CodedInputStream input, 6753 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6754 throws com.google.protobuf.InvalidProtocolBufferException { 6755 return new AssignRegionResponse(input, extensionRegistry); 6756 } 6757 }; 6758 6759 @java.lang.Override getParserForType()6760 public com.google.protobuf.Parser<AssignRegionResponse> getParserForType() { 6761 return PARSER; 6762 } 6763 initFields()6764 private void initFields() { 6765 } 6766 private byte memoizedIsInitialized = -1; isInitialized()6767 public final boolean isInitialized() { 6768 byte isInitialized = memoizedIsInitialized; 6769 if (isInitialized != -1) return isInitialized == 1; 6770 6771 memoizedIsInitialized = 1; 6772 return true; 6773 } 6774 writeTo(com.google.protobuf.CodedOutputStream output)6775 public void writeTo(com.google.protobuf.CodedOutputStream output) 6776 throws java.io.IOException { 6777 getSerializedSize(); 6778 getUnknownFields().writeTo(output); 6779 } 6780 6781 private int memoizedSerializedSize = -1; getSerializedSize()6782 public int getSerializedSize() { 6783 int size = memoizedSerializedSize; 6784 if (size != -1) return size; 6785 6786 size = 0; 6787 size += getUnknownFields().getSerializedSize(); 6788 memoizedSerializedSize = size; 6789 return size; 6790 } 6791 6792 private static final long serialVersionUID = 0L; 6793 @java.lang.Override writeReplace()6794 protected java.lang.Object writeReplace() 6795 throws java.io.ObjectStreamException { 6796 return super.writeReplace(); 6797 } 6798 6799 @java.lang.Override equals(final java.lang.Object obj)6800 public boolean equals(final java.lang.Object obj) { 6801 if (obj == this) { 6802 return true; 6803 } 6804 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse)) { 6805 return super.equals(obj); 6806 } 6807 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse) obj; 6808 6809 boolean result = true; 6810 result = result && 6811 getUnknownFields().equals(other.getUnknownFields()); 6812 return result; 6813 } 6814 6815 private int memoizedHashCode = 0; 6816 @java.lang.Override hashCode()6817 public int hashCode() { 6818 if (memoizedHashCode != 0) { 6819 return memoizedHashCode; 6820 } 6821 int hash = 41; 6822 hash = (19 * hash) + getDescriptorForType().hashCode(); 6823 hash = (29 * hash) + getUnknownFields().hashCode(); 6824 memoizedHashCode = hash; 6825 return hash; 6826 } 6827 parseFrom( com.google.protobuf.ByteString data)6828 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6829 com.google.protobuf.ByteString data) 6830 throws com.google.protobuf.InvalidProtocolBufferException { 6831 return PARSER.parseFrom(data); 6832 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6833 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6834 com.google.protobuf.ByteString data, 6835 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6836 throws com.google.protobuf.InvalidProtocolBufferException { 6837 return PARSER.parseFrom(data, extensionRegistry); 6838 } parseFrom(byte[] data)6839 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom(byte[] data) 6840 throws com.google.protobuf.InvalidProtocolBufferException { 6841 return PARSER.parseFrom(data); 6842 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6843 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6844 byte[] data, 6845 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6846 throws com.google.protobuf.InvalidProtocolBufferException { 6847 return PARSER.parseFrom(data, extensionRegistry); 6848 } parseFrom(java.io.InputStream input)6849 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom(java.io.InputStream input) 6850 throws java.io.IOException { 6851 return PARSER.parseFrom(input); 6852 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6853 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6854 java.io.InputStream input, 6855 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6856 throws java.io.IOException { 6857 return PARSER.parseFrom(input, extensionRegistry); 6858 } parseDelimitedFrom(java.io.InputStream input)6859 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseDelimitedFrom(java.io.InputStream input) 6860 throws java.io.IOException { 6861 return PARSER.parseDelimitedFrom(input); 6862 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6863 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseDelimitedFrom( 6864 java.io.InputStream input, 6865 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6866 throws java.io.IOException { 6867 return PARSER.parseDelimitedFrom(input, extensionRegistry); 6868 } parseFrom( com.google.protobuf.CodedInputStream input)6869 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6870 com.google.protobuf.CodedInputStream input) 6871 throws java.io.IOException { 6872 return PARSER.parseFrom(input); 6873 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6874 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( 6875 com.google.protobuf.CodedInputStream input, 6876 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6877 throws java.io.IOException { 6878 return PARSER.parseFrom(input, extensionRegistry); 6879 } 6880 newBuilder()6881 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()6882 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse prototype)6883 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse prototype) { 6884 return newBuilder().mergeFrom(prototype); 6885 } toBuilder()6886 public Builder toBuilder() { return newBuilder(this); } 6887 6888 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)6889 protected Builder newBuilderForType( 6890 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6891 Builder builder = new Builder(parent); 6892 return builder; 6893 } 6894 /** 6895 * Protobuf type {@code AssignRegionResponse} 6896 */ 6897 public static final class Builder extends 6898 com.google.protobuf.GeneratedMessage.Builder<Builder> 6899 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponseOrBuilder { 6900 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()6901 getDescriptor() { 6902 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionResponse_descriptor; 6903 } 6904 6905 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()6906 internalGetFieldAccessorTable() { 6907 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionResponse_fieldAccessorTable 6908 .ensureFieldAccessorsInitialized( 6909 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.Builder.class); 6910 } 6911 6912 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.newBuilder() Builder()6913 private Builder() { 6914 maybeForceBuilderInitialization(); 6915 } 6916 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)6917 private Builder( 6918 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 6919 super(parent); 6920 maybeForceBuilderInitialization(); 6921 } maybeForceBuilderInitialization()6922 private void maybeForceBuilderInitialization() { 6923 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 6924 } 6925 } create()6926 private static Builder create() { 6927 return new Builder(); 6928 } 6929 clear()6930 public Builder clear() { 6931 super.clear(); 6932 return this; 6933 } 6934 clone()6935 public Builder clone() { 6936 return create().mergeFrom(buildPartial()); 6937 } 6938 6939 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()6940 getDescriptorForType() { 6941 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_AssignRegionResponse_descriptor; 6942 } 6943 getDefaultInstanceForType()6944 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse getDefaultInstanceForType() { 6945 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); 6946 } 6947 build()6948 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse build() { 6949 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse result = buildPartial(); 6950 if (!result.isInitialized()) { 6951 throw newUninitializedMessageException(result); 6952 } 6953 return result; 6954 } 6955 buildPartial()6956 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse buildPartial() { 6957 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse(this); 6958 onBuilt(); 6959 return result; 6960 } 6961 mergeFrom(com.google.protobuf.Message other)6962 public Builder mergeFrom(com.google.protobuf.Message other) { 6963 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse) { 6964 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse)other); 6965 } else { 6966 super.mergeFrom(other); 6967 return this; 6968 } 6969 } 6970 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse other)6971 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse other) { 6972 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance()) return this; 6973 this.mergeUnknownFields(other.getUnknownFields()); 6974 return this; 6975 } 6976 isInitialized()6977 public final boolean isInitialized() { 6978 return true; 6979 } 6980 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)6981 public Builder mergeFrom( 6982 com.google.protobuf.CodedInputStream input, 6983 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 6984 throws java.io.IOException { 6985 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse parsedMessage = null; 6986 try { 6987 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 6988 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 6989 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.AssignRegionResponse) e.getUnfinishedMessage(); 6990 throw e; 6991 } finally { 6992 if (parsedMessage != null) { 6993 mergeFrom(parsedMessage); 6994 } 6995 } 6996 return this; 6997 } 6998 6999 // @@protoc_insertion_point(builder_scope:AssignRegionResponse) 7000 } 7001 7002 static { 7003 defaultInstance = new AssignRegionResponse(true); defaultInstance.initFields()7004 defaultInstance.initFields(); 7005 } 7006 7007 // @@protoc_insertion_point(class_scope:AssignRegionResponse) 7008 } 7009 7010 public interface UnassignRegionRequestOrBuilder 7011 extends com.google.protobuf.MessageOrBuilder { 7012 7013 // required .RegionSpecifier region = 1; 7014 /** 7015 * <code>required .RegionSpecifier region = 1;</code> 7016 */ hasRegion()7017 boolean hasRegion(); 7018 /** 7019 * <code>required .RegionSpecifier region = 1;</code> 7020 */ getRegion()7021 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); 7022 /** 7023 * <code>required .RegionSpecifier region = 1;</code> 7024 */ getRegionOrBuilder()7025 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); 7026 7027 // optional bool force = 2 [default = false]; 7028 /** 7029 * <code>optional bool force = 2 [default = false];</code> 7030 */ hasForce()7031 boolean hasForce(); 7032 /** 7033 * <code>optional bool force = 2 [default = false];</code> 7034 */ getForce()7035 boolean getForce(); 7036 } 7037 /** 7038 * Protobuf type {@code UnassignRegionRequest} 7039 */ 7040 public static final class UnassignRegionRequest extends 7041 com.google.protobuf.GeneratedMessage 7042 implements UnassignRegionRequestOrBuilder { 7043 // Use UnassignRegionRequest.newBuilder() to construct. UnassignRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)7044 private UnassignRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7045 super(builder); 7046 this.unknownFields = builder.getUnknownFields(); 7047 } UnassignRegionRequest(boolean noInit)7048 private UnassignRegionRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7049 7050 private static final UnassignRegionRequest defaultInstance; getDefaultInstance()7051 public static UnassignRegionRequest getDefaultInstance() { 7052 return defaultInstance; 7053 } 7054 getDefaultInstanceForType()7055 public UnassignRegionRequest getDefaultInstanceForType() { 7056 return defaultInstance; 7057 } 7058 7059 private final com.google.protobuf.UnknownFieldSet unknownFields; 7060 @java.lang.Override 7061 public final com.google.protobuf.UnknownFieldSet getUnknownFields()7062 getUnknownFields() { 7063 return this.unknownFields; 7064 } UnassignRegionRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7065 private UnassignRegionRequest( 7066 com.google.protobuf.CodedInputStream input, 7067 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7068 throws com.google.protobuf.InvalidProtocolBufferException { 7069 initFields(); 7070 int mutable_bitField0_ = 0; 7071 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7072 com.google.protobuf.UnknownFieldSet.newBuilder(); 7073 try { 7074 boolean done = false; 7075 while (!done) { 7076 int tag = input.readTag(); 7077 switch (tag) { 7078 case 0: 7079 done = true; 7080 break; 7081 default: { 7082 if (!parseUnknownField(input, unknownFields, 7083 extensionRegistry, tag)) { 7084 done = true; 7085 } 7086 break; 7087 } 7088 case 10: { 7089 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 7090 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7091 subBuilder = region_.toBuilder(); 7092 } 7093 region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 7094 if (subBuilder != null) { 7095 subBuilder.mergeFrom(region_); 7096 region_ = subBuilder.buildPartial(); 7097 } 7098 bitField0_ |= 0x00000001; 7099 break; 7100 } 7101 case 16: { 7102 bitField0_ |= 0x00000002; 7103 force_ = input.readBool(); 7104 break; 7105 } 7106 } 7107 } 7108 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7109 throw e.setUnfinishedMessage(this); 7110 } catch (java.io.IOException e) { 7111 throw new com.google.protobuf.InvalidProtocolBufferException( 7112 e.getMessage()).setUnfinishedMessage(this); 7113 } finally { 7114 this.unknownFields = unknownFields.build(); 7115 makeExtensionsImmutable(); 7116 } 7117 } 7118 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()7119 getDescriptor() { 7120 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionRequest_descriptor; 7121 } 7122 7123 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()7124 internalGetFieldAccessorTable() { 7125 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionRequest_fieldAccessorTable 7126 .ensureFieldAccessorsInitialized( 7127 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.Builder.class); 7128 } 7129 7130 public static com.google.protobuf.Parser<UnassignRegionRequest> PARSER = 7131 new com.google.protobuf.AbstractParser<UnassignRegionRequest>() { 7132 public UnassignRegionRequest parsePartialFrom( 7133 com.google.protobuf.CodedInputStream input, 7134 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7135 throws com.google.protobuf.InvalidProtocolBufferException { 7136 return new UnassignRegionRequest(input, extensionRegistry); 7137 } 7138 }; 7139 7140 @java.lang.Override getParserForType()7141 public com.google.protobuf.Parser<UnassignRegionRequest> getParserForType() { 7142 return PARSER; 7143 } 7144 7145 private int bitField0_; 7146 // required .RegionSpecifier region = 1; 7147 public static final int REGION_FIELD_NUMBER = 1; 7148 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_; 7149 /** 7150 * <code>required .RegionSpecifier region = 1;</code> 7151 */ hasRegion()7152 public boolean hasRegion() { 7153 return ((bitField0_ & 0x00000001) == 0x00000001); 7154 } 7155 /** 7156 * <code>required .RegionSpecifier region = 1;</code> 7157 */ getRegion()7158 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 7159 return region_; 7160 } 7161 /** 7162 * <code>required .RegionSpecifier region = 1;</code> 7163 */ getRegionOrBuilder()7164 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 7165 return region_; 7166 } 7167 7168 // optional bool force = 2 [default = false]; 7169 public static final int FORCE_FIELD_NUMBER = 2; 7170 private boolean force_; 7171 /** 7172 * <code>optional bool force = 2 [default = false];</code> 7173 */ hasForce()7174 public boolean hasForce() { 7175 return ((bitField0_ & 0x00000002) == 0x00000002); 7176 } 7177 /** 7178 * <code>optional bool force = 2 [default = false];</code> 7179 */ getForce()7180 public boolean getForce() { 7181 return force_; 7182 } 7183 initFields()7184 private void initFields() { 7185 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 7186 force_ = false; 7187 } 7188 private byte memoizedIsInitialized = -1; isInitialized()7189 public final boolean isInitialized() { 7190 byte isInitialized = memoizedIsInitialized; 7191 if (isInitialized != -1) return isInitialized == 1; 7192 7193 if (!hasRegion()) { 7194 memoizedIsInitialized = 0; 7195 return false; 7196 } 7197 if (!getRegion().isInitialized()) { 7198 memoizedIsInitialized = 0; 7199 return false; 7200 } 7201 memoizedIsInitialized = 1; 7202 return true; 7203 } 7204 writeTo(com.google.protobuf.CodedOutputStream output)7205 public void writeTo(com.google.protobuf.CodedOutputStream output) 7206 throws java.io.IOException { 7207 getSerializedSize(); 7208 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7209 output.writeMessage(1, region_); 7210 } 7211 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7212 output.writeBool(2, force_); 7213 } 7214 getUnknownFields().writeTo(output); 7215 } 7216 7217 private int memoizedSerializedSize = -1; getSerializedSize()7218 public int getSerializedSize() { 7219 int size = memoizedSerializedSize; 7220 if (size != -1) return size; 7221 7222 size = 0; 7223 if (((bitField0_ & 0x00000001) == 0x00000001)) { 7224 size += com.google.protobuf.CodedOutputStream 7225 .computeMessageSize(1, region_); 7226 } 7227 if (((bitField0_ & 0x00000002) == 0x00000002)) { 7228 size += com.google.protobuf.CodedOutputStream 7229 .computeBoolSize(2, force_); 7230 } 7231 size += getUnknownFields().getSerializedSize(); 7232 memoizedSerializedSize = size; 7233 return size; 7234 } 7235 7236 private static final long serialVersionUID = 0L; 7237 @java.lang.Override writeReplace()7238 protected java.lang.Object writeReplace() 7239 throws java.io.ObjectStreamException { 7240 return super.writeReplace(); 7241 } 7242 7243 @java.lang.Override equals(final java.lang.Object obj)7244 public boolean equals(final java.lang.Object obj) { 7245 if (obj == this) { 7246 return true; 7247 } 7248 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest)) { 7249 return super.equals(obj); 7250 } 7251 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest) obj; 7252 7253 boolean result = true; 7254 result = result && (hasRegion() == other.hasRegion()); 7255 if (hasRegion()) { 7256 result = result && getRegion() 7257 .equals(other.getRegion()); 7258 } 7259 result = result && (hasForce() == other.hasForce()); 7260 if (hasForce()) { 7261 result = result && (getForce() 7262 == other.getForce()); 7263 } 7264 result = result && 7265 getUnknownFields().equals(other.getUnknownFields()); 7266 return result; 7267 } 7268 7269 private int memoizedHashCode = 0; 7270 @java.lang.Override hashCode()7271 public int hashCode() { 7272 if (memoizedHashCode != 0) { 7273 return memoizedHashCode; 7274 } 7275 int hash = 41; 7276 hash = (19 * hash) + getDescriptorForType().hashCode(); 7277 if (hasRegion()) { 7278 hash = (37 * hash) + REGION_FIELD_NUMBER; 7279 hash = (53 * hash) + getRegion().hashCode(); 7280 } 7281 if (hasForce()) { 7282 hash = (37 * hash) + FORCE_FIELD_NUMBER; 7283 hash = (53 * hash) + hashBoolean(getForce()); 7284 } 7285 hash = (29 * hash) + getUnknownFields().hashCode(); 7286 memoizedHashCode = hash; 7287 return hash; 7288 } 7289 parseFrom( com.google.protobuf.ByteString data)7290 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7291 com.google.protobuf.ByteString data) 7292 throws com.google.protobuf.InvalidProtocolBufferException { 7293 return PARSER.parseFrom(data); 7294 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7295 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7296 com.google.protobuf.ByteString data, 7297 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7298 throws com.google.protobuf.InvalidProtocolBufferException { 7299 return PARSER.parseFrom(data, extensionRegistry); 7300 } parseFrom(byte[] data)7301 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom(byte[] data) 7302 throws com.google.protobuf.InvalidProtocolBufferException { 7303 return PARSER.parseFrom(data); 7304 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7305 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7306 byte[] data, 7307 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7308 throws com.google.protobuf.InvalidProtocolBufferException { 7309 return PARSER.parseFrom(data, extensionRegistry); 7310 } parseFrom(java.io.InputStream input)7311 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom(java.io.InputStream input) 7312 throws java.io.IOException { 7313 return PARSER.parseFrom(input); 7314 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7315 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7316 java.io.InputStream input, 7317 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7318 throws java.io.IOException { 7319 return PARSER.parseFrom(input, extensionRegistry); 7320 } parseDelimitedFrom(java.io.InputStream input)7321 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseDelimitedFrom(java.io.InputStream input) 7322 throws java.io.IOException { 7323 return PARSER.parseDelimitedFrom(input); 7324 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7325 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseDelimitedFrom( 7326 java.io.InputStream input, 7327 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7328 throws java.io.IOException { 7329 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7330 } parseFrom( com.google.protobuf.CodedInputStream input)7331 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7332 com.google.protobuf.CodedInputStream input) 7333 throws java.io.IOException { 7334 return PARSER.parseFrom(input); 7335 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7336 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( 7337 com.google.protobuf.CodedInputStream input, 7338 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7339 throws java.io.IOException { 7340 return PARSER.parseFrom(input, extensionRegistry); 7341 } 7342 newBuilder()7343 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()7344 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest prototype)7345 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest prototype) { 7346 return newBuilder().mergeFrom(prototype); 7347 } toBuilder()7348 public Builder toBuilder() { return newBuilder(this); } 7349 7350 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)7351 protected Builder newBuilderForType( 7352 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7353 Builder builder = new Builder(parent); 7354 return builder; 7355 } 7356 /** 7357 * Protobuf type {@code UnassignRegionRequest} 7358 */ 7359 public static final class Builder extends 7360 com.google.protobuf.GeneratedMessage.Builder<Builder> 7361 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequestOrBuilder { 7362 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()7363 getDescriptor() { 7364 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionRequest_descriptor; 7365 } 7366 7367 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()7368 internalGetFieldAccessorTable() { 7369 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionRequest_fieldAccessorTable 7370 .ensureFieldAccessorsInitialized( 7371 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.Builder.class); 7372 } 7373 7374 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.newBuilder() Builder()7375 private Builder() { 7376 maybeForceBuilderInitialization(); 7377 } 7378 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)7379 private Builder( 7380 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7381 super(parent); 7382 maybeForceBuilderInitialization(); 7383 } maybeForceBuilderInitialization()7384 private void maybeForceBuilderInitialization() { 7385 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7386 getRegionFieldBuilder(); 7387 } 7388 } create()7389 private static Builder create() { 7390 return new Builder(); 7391 } 7392 clear()7393 public Builder clear() { 7394 super.clear(); 7395 if (regionBuilder_ == null) { 7396 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 7397 } else { 7398 regionBuilder_.clear(); 7399 } 7400 bitField0_ = (bitField0_ & ~0x00000001); 7401 force_ = false; 7402 bitField0_ = (bitField0_ & ~0x00000002); 7403 return this; 7404 } 7405 clone()7406 public Builder clone() { 7407 return create().mergeFrom(buildPartial()); 7408 } 7409 7410 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()7411 getDescriptorForType() { 7412 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionRequest_descriptor; 7413 } 7414 getDefaultInstanceForType()7415 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest getDefaultInstanceForType() { 7416 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); 7417 } 7418 build()7419 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest build() { 7420 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest result = buildPartial(); 7421 if (!result.isInitialized()) { 7422 throw newUninitializedMessageException(result); 7423 } 7424 return result; 7425 } 7426 buildPartial()7427 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest buildPartial() { 7428 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest(this); 7429 int from_bitField0_ = bitField0_; 7430 int to_bitField0_ = 0; 7431 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 7432 to_bitField0_ |= 0x00000001; 7433 } 7434 if (regionBuilder_ == null) { 7435 result.region_ = region_; 7436 } else { 7437 result.region_ = regionBuilder_.build(); 7438 } 7439 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 7440 to_bitField0_ |= 0x00000002; 7441 } 7442 result.force_ = force_; 7443 result.bitField0_ = to_bitField0_; 7444 onBuilt(); 7445 return result; 7446 } 7447 mergeFrom(com.google.protobuf.Message other)7448 public Builder mergeFrom(com.google.protobuf.Message other) { 7449 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest) { 7450 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest)other); 7451 } else { 7452 super.mergeFrom(other); 7453 return this; 7454 } 7455 } 7456 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest other)7457 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest other) { 7458 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance()) return this; 7459 if (other.hasRegion()) { 7460 mergeRegion(other.getRegion()); 7461 } 7462 if (other.hasForce()) { 7463 setForce(other.getForce()); 7464 } 7465 this.mergeUnknownFields(other.getUnknownFields()); 7466 return this; 7467 } 7468 isInitialized()7469 public final boolean isInitialized() { 7470 if (!hasRegion()) { 7471 7472 return false; 7473 } 7474 if (!getRegion().isInitialized()) { 7475 7476 return false; 7477 } 7478 return true; 7479 } 7480 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7481 public Builder mergeFrom( 7482 com.google.protobuf.CodedInputStream input, 7483 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7484 throws java.io.IOException { 7485 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest parsedMessage = null; 7486 try { 7487 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7488 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7489 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionRequest) e.getUnfinishedMessage(); 7490 throw e; 7491 } finally { 7492 if (parsedMessage != null) { 7493 mergeFrom(parsedMessage); 7494 } 7495 } 7496 return this; 7497 } 7498 private int bitField0_; 7499 7500 // required .RegionSpecifier region = 1; 7501 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 7502 private com.google.protobuf.SingleFieldBuilder< 7503 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; 7504 /** 7505 * <code>required .RegionSpecifier region = 1;</code> 7506 */ hasRegion()7507 public boolean hasRegion() { 7508 return ((bitField0_ & 0x00000001) == 0x00000001); 7509 } 7510 /** 7511 * <code>required .RegionSpecifier region = 1;</code> 7512 */ getRegion()7513 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 7514 if (regionBuilder_ == null) { 7515 return region_; 7516 } else { 7517 return regionBuilder_.getMessage(); 7518 } 7519 } 7520 /** 7521 * <code>required .RegionSpecifier region = 1;</code> 7522 */ setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)7523 public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 7524 if (regionBuilder_ == null) { 7525 if (value == null) { 7526 throw new NullPointerException(); 7527 } 7528 region_ = value; 7529 onChanged(); 7530 } else { 7531 regionBuilder_.setMessage(value); 7532 } 7533 bitField0_ |= 0x00000001; 7534 return this; 7535 } 7536 /** 7537 * <code>required .RegionSpecifier region = 1;</code> 7538 */ setRegion( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)7539 public Builder setRegion( 7540 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 7541 if (regionBuilder_ == null) { 7542 region_ = builderForValue.build(); 7543 onChanged(); 7544 } else { 7545 regionBuilder_.setMessage(builderForValue.build()); 7546 } 7547 bitField0_ |= 0x00000001; 7548 return this; 7549 } 7550 /** 7551 * <code>required .RegionSpecifier region = 1;</code> 7552 */ mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)7553 public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 7554 if (regionBuilder_ == null) { 7555 if (((bitField0_ & 0x00000001) == 0x00000001) && 7556 region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 7557 region_ = 7558 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); 7559 } else { 7560 region_ = value; 7561 } 7562 onChanged(); 7563 } else { 7564 regionBuilder_.mergeFrom(value); 7565 } 7566 bitField0_ |= 0x00000001; 7567 return this; 7568 } 7569 /** 7570 * <code>required .RegionSpecifier region = 1;</code> 7571 */ clearRegion()7572 public Builder clearRegion() { 7573 if (regionBuilder_ == null) { 7574 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 7575 onChanged(); 7576 } else { 7577 regionBuilder_.clear(); 7578 } 7579 bitField0_ = (bitField0_ & ~0x00000001); 7580 return this; 7581 } 7582 /** 7583 * <code>required .RegionSpecifier region = 1;</code> 7584 */ getRegionBuilder()7585 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { 7586 bitField0_ |= 0x00000001; 7587 onChanged(); 7588 return getRegionFieldBuilder().getBuilder(); 7589 } 7590 /** 7591 * <code>required .RegionSpecifier region = 1;</code> 7592 */ getRegionOrBuilder()7593 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 7594 if (regionBuilder_ != null) { 7595 return regionBuilder_.getMessageOrBuilder(); 7596 } else { 7597 return region_; 7598 } 7599 } 7600 /** 7601 * <code>required .RegionSpecifier region = 1;</code> 7602 */ 7603 private com.google.protobuf.SingleFieldBuilder< 7604 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder()7605 getRegionFieldBuilder() { 7606 if (regionBuilder_ == null) { 7607 regionBuilder_ = new com.google.protobuf.SingleFieldBuilder< 7608 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 7609 region_, 7610 getParentForChildren(), 7611 isClean()); 7612 region_ = null; 7613 } 7614 return regionBuilder_; 7615 } 7616 7617 // optional bool force = 2 [default = false]; 7618 private boolean force_ ; 7619 /** 7620 * <code>optional bool force = 2 [default = false];</code> 7621 */ hasForce()7622 public boolean hasForce() { 7623 return ((bitField0_ & 0x00000002) == 0x00000002); 7624 } 7625 /** 7626 * <code>optional bool force = 2 [default = false];</code> 7627 */ getForce()7628 public boolean getForce() { 7629 return force_; 7630 } 7631 /** 7632 * <code>optional bool force = 2 [default = false];</code> 7633 */ setForce(boolean value)7634 public Builder setForce(boolean value) { 7635 bitField0_ |= 0x00000002; 7636 force_ = value; 7637 onChanged(); 7638 return this; 7639 } 7640 /** 7641 * <code>optional bool force = 2 [default = false];</code> 7642 */ clearForce()7643 public Builder clearForce() { 7644 bitField0_ = (bitField0_ & ~0x00000002); 7645 force_ = false; 7646 onChanged(); 7647 return this; 7648 } 7649 7650 // @@protoc_insertion_point(builder_scope:UnassignRegionRequest) 7651 } 7652 7653 static { 7654 defaultInstance = new UnassignRegionRequest(true); defaultInstance.initFields()7655 defaultInstance.initFields(); 7656 } 7657 7658 // @@protoc_insertion_point(class_scope:UnassignRegionRequest) 7659 } 7660 7661 public interface UnassignRegionResponseOrBuilder 7662 extends com.google.protobuf.MessageOrBuilder { 7663 } 7664 /** 7665 * Protobuf type {@code UnassignRegionResponse} 7666 */ 7667 public static final class UnassignRegionResponse extends 7668 com.google.protobuf.GeneratedMessage 7669 implements UnassignRegionResponseOrBuilder { 7670 // Use UnassignRegionResponse.newBuilder() to construct. UnassignRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)7671 private UnassignRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 7672 super(builder); 7673 this.unknownFields = builder.getUnknownFields(); 7674 } UnassignRegionResponse(boolean noInit)7675 private UnassignRegionResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 7676 7677 private static final UnassignRegionResponse defaultInstance; getDefaultInstance()7678 public static UnassignRegionResponse getDefaultInstance() { 7679 return defaultInstance; 7680 } 7681 getDefaultInstanceForType()7682 public UnassignRegionResponse getDefaultInstanceForType() { 7683 return defaultInstance; 7684 } 7685 7686 private final com.google.protobuf.UnknownFieldSet unknownFields; 7687 @java.lang.Override 7688 public final com.google.protobuf.UnknownFieldSet getUnknownFields()7689 getUnknownFields() { 7690 return this.unknownFields; 7691 } UnassignRegionResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7692 private UnassignRegionResponse( 7693 com.google.protobuf.CodedInputStream input, 7694 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7695 throws com.google.protobuf.InvalidProtocolBufferException { 7696 initFields(); 7697 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 7698 com.google.protobuf.UnknownFieldSet.newBuilder(); 7699 try { 7700 boolean done = false; 7701 while (!done) { 7702 int tag = input.readTag(); 7703 switch (tag) { 7704 case 0: 7705 done = true; 7706 break; 7707 default: { 7708 if (!parseUnknownField(input, unknownFields, 7709 extensionRegistry, tag)) { 7710 done = true; 7711 } 7712 break; 7713 } 7714 } 7715 } 7716 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7717 throw e.setUnfinishedMessage(this); 7718 } catch (java.io.IOException e) { 7719 throw new com.google.protobuf.InvalidProtocolBufferException( 7720 e.getMessage()).setUnfinishedMessage(this); 7721 } finally { 7722 this.unknownFields = unknownFields.build(); 7723 makeExtensionsImmutable(); 7724 } 7725 } 7726 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()7727 getDescriptor() { 7728 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionResponse_descriptor; 7729 } 7730 7731 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()7732 internalGetFieldAccessorTable() { 7733 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionResponse_fieldAccessorTable 7734 .ensureFieldAccessorsInitialized( 7735 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.Builder.class); 7736 } 7737 7738 public static com.google.protobuf.Parser<UnassignRegionResponse> PARSER = 7739 new com.google.protobuf.AbstractParser<UnassignRegionResponse>() { 7740 public UnassignRegionResponse parsePartialFrom( 7741 com.google.protobuf.CodedInputStream input, 7742 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7743 throws com.google.protobuf.InvalidProtocolBufferException { 7744 return new UnassignRegionResponse(input, extensionRegistry); 7745 } 7746 }; 7747 7748 @java.lang.Override getParserForType()7749 public com.google.protobuf.Parser<UnassignRegionResponse> getParserForType() { 7750 return PARSER; 7751 } 7752 initFields()7753 private void initFields() { 7754 } 7755 private byte memoizedIsInitialized = -1; isInitialized()7756 public final boolean isInitialized() { 7757 byte isInitialized = memoizedIsInitialized; 7758 if (isInitialized != -1) return isInitialized == 1; 7759 7760 memoizedIsInitialized = 1; 7761 return true; 7762 } 7763 writeTo(com.google.protobuf.CodedOutputStream output)7764 public void writeTo(com.google.protobuf.CodedOutputStream output) 7765 throws java.io.IOException { 7766 getSerializedSize(); 7767 getUnknownFields().writeTo(output); 7768 } 7769 7770 private int memoizedSerializedSize = -1; getSerializedSize()7771 public int getSerializedSize() { 7772 int size = memoizedSerializedSize; 7773 if (size != -1) return size; 7774 7775 size = 0; 7776 size += getUnknownFields().getSerializedSize(); 7777 memoizedSerializedSize = size; 7778 return size; 7779 } 7780 7781 private static final long serialVersionUID = 0L; 7782 @java.lang.Override writeReplace()7783 protected java.lang.Object writeReplace() 7784 throws java.io.ObjectStreamException { 7785 return super.writeReplace(); 7786 } 7787 7788 @java.lang.Override equals(final java.lang.Object obj)7789 public boolean equals(final java.lang.Object obj) { 7790 if (obj == this) { 7791 return true; 7792 } 7793 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse)) { 7794 return super.equals(obj); 7795 } 7796 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse) obj; 7797 7798 boolean result = true; 7799 result = result && 7800 getUnknownFields().equals(other.getUnknownFields()); 7801 return result; 7802 } 7803 7804 private int memoizedHashCode = 0; 7805 @java.lang.Override hashCode()7806 public int hashCode() { 7807 if (memoizedHashCode != 0) { 7808 return memoizedHashCode; 7809 } 7810 int hash = 41; 7811 hash = (19 * hash) + getDescriptorForType().hashCode(); 7812 hash = (29 * hash) + getUnknownFields().hashCode(); 7813 memoizedHashCode = hash; 7814 return hash; 7815 } 7816 parseFrom( com.google.protobuf.ByteString data)7817 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7818 com.google.protobuf.ByteString data) 7819 throws com.google.protobuf.InvalidProtocolBufferException { 7820 return PARSER.parseFrom(data); 7821 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7822 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7823 com.google.protobuf.ByteString data, 7824 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7825 throws com.google.protobuf.InvalidProtocolBufferException { 7826 return PARSER.parseFrom(data, extensionRegistry); 7827 } parseFrom(byte[] data)7828 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom(byte[] data) 7829 throws com.google.protobuf.InvalidProtocolBufferException { 7830 return PARSER.parseFrom(data); 7831 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7832 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7833 byte[] data, 7834 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7835 throws com.google.protobuf.InvalidProtocolBufferException { 7836 return PARSER.parseFrom(data, extensionRegistry); 7837 } parseFrom(java.io.InputStream input)7838 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom(java.io.InputStream input) 7839 throws java.io.IOException { 7840 return PARSER.parseFrom(input); 7841 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7842 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7843 java.io.InputStream input, 7844 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7845 throws java.io.IOException { 7846 return PARSER.parseFrom(input, extensionRegistry); 7847 } parseDelimitedFrom(java.io.InputStream input)7848 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseDelimitedFrom(java.io.InputStream input) 7849 throws java.io.IOException { 7850 return PARSER.parseDelimitedFrom(input); 7851 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7852 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseDelimitedFrom( 7853 java.io.InputStream input, 7854 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7855 throws java.io.IOException { 7856 return PARSER.parseDelimitedFrom(input, extensionRegistry); 7857 } parseFrom( com.google.protobuf.CodedInputStream input)7858 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7859 com.google.protobuf.CodedInputStream input) 7860 throws java.io.IOException { 7861 return PARSER.parseFrom(input); 7862 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7863 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( 7864 com.google.protobuf.CodedInputStream input, 7865 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7866 throws java.io.IOException { 7867 return PARSER.parseFrom(input, extensionRegistry); 7868 } 7869 newBuilder()7870 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()7871 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse prototype)7872 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse prototype) { 7873 return newBuilder().mergeFrom(prototype); 7874 } toBuilder()7875 public Builder toBuilder() { return newBuilder(this); } 7876 7877 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)7878 protected Builder newBuilderForType( 7879 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7880 Builder builder = new Builder(parent); 7881 return builder; 7882 } 7883 /** 7884 * Protobuf type {@code UnassignRegionResponse} 7885 */ 7886 public static final class Builder extends 7887 com.google.protobuf.GeneratedMessage.Builder<Builder> 7888 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponseOrBuilder { 7889 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()7890 getDescriptor() { 7891 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionResponse_descriptor; 7892 } 7893 7894 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()7895 internalGetFieldAccessorTable() { 7896 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionResponse_fieldAccessorTable 7897 .ensureFieldAccessorsInitialized( 7898 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.Builder.class); 7899 } 7900 7901 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.newBuilder() Builder()7902 private Builder() { 7903 maybeForceBuilderInitialization(); 7904 } 7905 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)7906 private Builder( 7907 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 7908 super(parent); 7909 maybeForceBuilderInitialization(); 7910 } maybeForceBuilderInitialization()7911 private void maybeForceBuilderInitialization() { 7912 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 7913 } 7914 } create()7915 private static Builder create() { 7916 return new Builder(); 7917 } 7918 clear()7919 public Builder clear() { 7920 super.clear(); 7921 return this; 7922 } 7923 clone()7924 public Builder clone() { 7925 return create().mergeFrom(buildPartial()); 7926 } 7927 7928 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()7929 getDescriptorForType() { 7930 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_UnassignRegionResponse_descriptor; 7931 } 7932 getDefaultInstanceForType()7933 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse getDefaultInstanceForType() { 7934 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); 7935 } 7936 build()7937 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse build() { 7938 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse result = buildPartial(); 7939 if (!result.isInitialized()) { 7940 throw newUninitializedMessageException(result); 7941 } 7942 return result; 7943 } 7944 buildPartial()7945 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse buildPartial() { 7946 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse(this); 7947 onBuilt(); 7948 return result; 7949 } 7950 mergeFrom(com.google.protobuf.Message other)7951 public Builder mergeFrom(com.google.protobuf.Message other) { 7952 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse) { 7953 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse)other); 7954 } else { 7955 super.mergeFrom(other); 7956 return this; 7957 } 7958 } 7959 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse other)7960 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse other) { 7961 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance()) return this; 7962 this.mergeUnknownFields(other.getUnknownFields()); 7963 return this; 7964 } 7965 isInitialized()7966 public final boolean isInitialized() { 7967 return true; 7968 } 7969 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)7970 public Builder mergeFrom( 7971 com.google.protobuf.CodedInputStream input, 7972 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 7973 throws java.io.IOException { 7974 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse parsedMessage = null; 7975 try { 7976 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 7977 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 7978 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.UnassignRegionResponse) e.getUnfinishedMessage(); 7979 throw e; 7980 } finally { 7981 if (parsedMessage != null) { 7982 mergeFrom(parsedMessage); 7983 } 7984 } 7985 return this; 7986 } 7987 7988 // @@protoc_insertion_point(builder_scope:UnassignRegionResponse) 7989 } 7990 7991 static { 7992 defaultInstance = new UnassignRegionResponse(true); defaultInstance.initFields()7993 defaultInstance.initFields(); 7994 } 7995 7996 // @@protoc_insertion_point(class_scope:UnassignRegionResponse) 7997 } 7998 7999 public interface OfflineRegionRequestOrBuilder 8000 extends com.google.protobuf.MessageOrBuilder { 8001 8002 // required .RegionSpecifier region = 1; 8003 /** 8004 * <code>required .RegionSpecifier region = 1;</code> 8005 */ hasRegion()8006 boolean hasRegion(); 8007 /** 8008 * <code>required .RegionSpecifier region = 1;</code> 8009 */ getRegion()8010 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); 8011 /** 8012 * <code>required .RegionSpecifier region = 1;</code> 8013 */ getRegionOrBuilder()8014 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); 8015 } 8016 /** 8017 * Protobuf type {@code OfflineRegionRequest} 8018 */ 8019 public static final class OfflineRegionRequest extends 8020 com.google.protobuf.GeneratedMessage 8021 implements OfflineRegionRequestOrBuilder { 8022 // Use OfflineRegionRequest.newBuilder() to construct. OfflineRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder)8023 private OfflineRegionRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8024 super(builder); 8025 this.unknownFields = builder.getUnknownFields(); 8026 } OfflineRegionRequest(boolean noInit)8027 private OfflineRegionRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8028 8029 private static final OfflineRegionRequest defaultInstance; getDefaultInstance()8030 public static OfflineRegionRequest getDefaultInstance() { 8031 return defaultInstance; 8032 } 8033 getDefaultInstanceForType()8034 public OfflineRegionRequest getDefaultInstanceForType() { 8035 return defaultInstance; 8036 } 8037 8038 private final com.google.protobuf.UnknownFieldSet unknownFields; 8039 @java.lang.Override 8040 public final com.google.protobuf.UnknownFieldSet getUnknownFields()8041 getUnknownFields() { 8042 return this.unknownFields; 8043 } OfflineRegionRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8044 private OfflineRegionRequest( 8045 com.google.protobuf.CodedInputStream input, 8046 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8047 throws com.google.protobuf.InvalidProtocolBufferException { 8048 initFields(); 8049 int mutable_bitField0_ = 0; 8050 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8051 com.google.protobuf.UnknownFieldSet.newBuilder(); 8052 try { 8053 boolean done = false; 8054 while (!done) { 8055 int tag = input.readTag(); 8056 switch (tag) { 8057 case 0: 8058 done = true; 8059 break; 8060 default: { 8061 if (!parseUnknownField(input, unknownFields, 8062 extensionRegistry, tag)) { 8063 done = true; 8064 } 8065 break; 8066 } 8067 case 10: { 8068 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; 8069 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8070 subBuilder = region_.toBuilder(); 8071 } 8072 region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); 8073 if (subBuilder != null) { 8074 subBuilder.mergeFrom(region_); 8075 region_ = subBuilder.buildPartial(); 8076 } 8077 bitField0_ |= 0x00000001; 8078 break; 8079 } 8080 } 8081 } 8082 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8083 throw e.setUnfinishedMessage(this); 8084 } catch (java.io.IOException e) { 8085 throw new com.google.protobuf.InvalidProtocolBufferException( 8086 e.getMessage()).setUnfinishedMessage(this); 8087 } finally { 8088 this.unknownFields = unknownFields.build(); 8089 makeExtensionsImmutable(); 8090 } 8091 } 8092 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()8093 getDescriptor() { 8094 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionRequest_descriptor; 8095 } 8096 8097 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()8098 internalGetFieldAccessorTable() { 8099 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionRequest_fieldAccessorTable 8100 .ensureFieldAccessorsInitialized( 8101 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.Builder.class); 8102 } 8103 8104 public static com.google.protobuf.Parser<OfflineRegionRequest> PARSER = 8105 new com.google.protobuf.AbstractParser<OfflineRegionRequest>() { 8106 public OfflineRegionRequest parsePartialFrom( 8107 com.google.protobuf.CodedInputStream input, 8108 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8109 throws com.google.protobuf.InvalidProtocolBufferException { 8110 return new OfflineRegionRequest(input, extensionRegistry); 8111 } 8112 }; 8113 8114 @java.lang.Override getParserForType()8115 public com.google.protobuf.Parser<OfflineRegionRequest> getParserForType() { 8116 return PARSER; 8117 } 8118 8119 private int bitField0_; 8120 // required .RegionSpecifier region = 1; 8121 public static final int REGION_FIELD_NUMBER = 1; 8122 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_; 8123 /** 8124 * <code>required .RegionSpecifier region = 1;</code> 8125 */ hasRegion()8126 public boolean hasRegion() { 8127 return ((bitField0_ & 0x00000001) == 0x00000001); 8128 } 8129 /** 8130 * <code>required .RegionSpecifier region = 1;</code> 8131 */ getRegion()8132 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 8133 return region_; 8134 } 8135 /** 8136 * <code>required .RegionSpecifier region = 1;</code> 8137 */ getRegionOrBuilder()8138 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 8139 return region_; 8140 } 8141 initFields()8142 private void initFields() { 8143 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 8144 } 8145 private byte memoizedIsInitialized = -1; isInitialized()8146 public final boolean isInitialized() { 8147 byte isInitialized = memoizedIsInitialized; 8148 if (isInitialized != -1) return isInitialized == 1; 8149 8150 if (!hasRegion()) { 8151 memoizedIsInitialized = 0; 8152 return false; 8153 } 8154 if (!getRegion().isInitialized()) { 8155 memoizedIsInitialized = 0; 8156 return false; 8157 } 8158 memoizedIsInitialized = 1; 8159 return true; 8160 } 8161 writeTo(com.google.protobuf.CodedOutputStream output)8162 public void writeTo(com.google.protobuf.CodedOutputStream output) 8163 throws java.io.IOException { 8164 getSerializedSize(); 8165 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8166 output.writeMessage(1, region_); 8167 } 8168 getUnknownFields().writeTo(output); 8169 } 8170 8171 private int memoizedSerializedSize = -1; getSerializedSize()8172 public int getSerializedSize() { 8173 int size = memoizedSerializedSize; 8174 if (size != -1) return size; 8175 8176 size = 0; 8177 if (((bitField0_ & 0x00000001) == 0x00000001)) { 8178 size += com.google.protobuf.CodedOutputStream 8179 .computeMessageSize(1, region_); 8180 } 8181 size += getUnknownFields().getSerializedSize(); 8182 memoizedSerializedSize = size; 8183 return size; 8184 } 8185 8186 private static final long serialVersionUID = 0L; 8187 @java.lang.Override writeReplace()8188 protected java.lang.Object writeReplace() 8189 throws java.io.ObjectStreamException { 8190 return super.writeReplace(); 8191 } 8192 8193 @java.lang.Override equals(final java.lang.Object obj)8194 public boolean equals(final java.lang.Object obj) { 8195 if (obj == this) { 8196 return true; 8197 } 8198 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest)) { 8199 return super.equals(obj); 8200 } 8201 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest) obj; 8202 8203 boolean result = true; 8204 result = result && (hasRegion() == other.hasRegion()); 8205 if (hasRegion()) { 8206 result = result && getRegion() 8207 .equals(other.getRegion()); 8208 } 8209 result = result && 8210 getUnknownFields().equals(other.getUnknownFields()); 8211 return result; 8212 } 8213 8214 private int memoizedHashCode = 0; 8215 @java.lang.Override hashCode()8216 public int hashCode() { 8217 if (memoizedHashCode != 0) { 8218 return memoizedHashCode; 8219 } 8220 int hash = 41; 8221 hash = (19 * hash) + getDescriptorForType().hashCode(); 8222 if (hasRegion()) { 8223 hash = (37 * hash) + REGION_FIELD_NUMBER; 8224 hash = (53 * hash) + getRegion().hashCode(); 8225 } 8226 hash = (29 * hash) + getUnknownFields().hashCode(); 8227 memoizedHashCode = hash; 8228 return hash; 8229 } 8230 parseFrom( com.google.protobuf.ByteString data)8231 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8232 com.google.protobuf.ByteString data) 8233 throws com.google.protobuf.InvalidProtocolBufferException { 8234 return PARSER.parseFrom(data); 8235 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8236 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8237 com.google.protobuf.ByteString data, 8238 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8239 throws com.google.protobuf.InvalidProtocolBufferException { 8240 return PARSER.parseFrom(data, extensionRegistry); 8241 } parseFrom(byte[] data)8242 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom(byte[] data) 8243 throws com.google.protobuf.InvalidProtocolBufferException { 8244 return PARSER.parseFrom(data); 8245 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8246 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8247 byte[] data, 8248 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8249 throws com.google.protobuf.InvalidProtocolBufferException { 8250 return PARSER.parseFrom(data, extensionRegistry); 8251 } parseFrom(java.io.InputStream input)8252 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom(java.io.InputStream input) 8253 throws java.io.IOException { 8254 return PARSER.parseFrom(input); 8255 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8256 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8257 java.io.InputStream input, 8258 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8259 throws java.io.IOException { 8260 return PARSER.parseFrom(input, extensionRegistry); 8261 } parseDelimitedFrom(java.io.InputStream input)8262 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseDelimitedFrom(java.io.InputStream input) 8263 throws java.io.IOException { 8264 return PARSER.parseDelimitedFrom(input); 8265 } parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8266 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseDelimitedFrom( 8267 java.io.InputStream input, 8268 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8269 throws java.io.IOException { 8270 return PARSER.parseDelimitedFrom(input, extensionRegistry); 8271 } parseFrom( com.google.protobuf.CodedInputStream input)8272 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8273 com.google.protobuf.CodedInputStream input) 8274 throws java.io.IOException { 8275 return PARSER.parseFrom(input); 8276 } parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8277 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( 8278 com.google.protobuf.CodedInputStream input, 8279 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8280 throws java.io.IOException { 8281 return PARSER.parseFrom(input, extensionRegistry); 8282 } 8283 newBuilder()8284 public static Builder newBuilder() { return Builder.create(); } newBuilderForType()8285 public Builder newBuilderForType() { return newBuilder(); } newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest prototype)8286 public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest prototype) { 8287 return newBuilder().mergeFrom(prototype); 8288 } toBuilder()8289 public Builder toBuilder() { return newBuilder(this); } 8290 8291 @java.lang.Override newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent)8292 protected Builder newBuilderForType( 8293 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8294 Builder builder = new Builder(parent); 8295 return builder; 8296 } 8297 /** 8298 * Protobuf type {@code OfflineRegionRequest} 8299 */ 8300 public static final class Builder extends 8301 com.google.protobuf.GeneratedMessage.Builder<Builder> 8302 implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequestOrBuilder { 8303 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()8304 getDescriptor() { 8305 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionRequest_descriptor; 8306 } 8307 8308 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()8309 internalGetFieldAccessorTable() { 8310 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionRequest_fieldAccessorTable 8311 .ensureFieldAccessorsInitialized( 8312 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.Builder.class); 8313 } 8314 8315 // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.newBuilder() Builder()8316 private Builder() { 8317 maybeForceBuilderInitialization(); 8318 } 8319 Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent)8320 private Builder( 8321 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 8322 super(parent); 8323 maybeForceBuilderInitialization(); 8324 } maybeForceBuilderInitialization()8325 private void maybeForceBuilderInitialization() { 8326 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 8327 getRegionFieldBuilder(); 8328 } 8329 } create()8330 private static Builder create() { 8331 return new Builder(); 8332 } 8333 clear()8334 public Builder clear() { 8335 super.clear(); 8336 if (regionBuilder_ == null) { 8337 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 8338 } else { 8339 regionBuilder_.clear(); 8340 } 8341 bitField0_ = (bitField0_ & ~0x00000001); 8342 return this; 8343 } 8344 clone()8345 public Builder clone() { 8346 return create().mergeFrom(buildPartial()); 8347 } 8348 8349 public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()8350 getDescriptorForType() { 8351 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionRequest_descriptor; 8352 } 8353 getDefaultInstanceForType()8354 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest getDefaultInstanceForType() { 8355 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); 8356 } 8357 build()8358 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest build() { 8359 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest result = buildPartial(); 8360 if (!result.isInitialized()) { 8361 throw newUninitializedMessageException(result); 8362 } 8363 return result; 8364 } 8365 buildPartial()8366 public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest buildPartial() { 8367 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest(this); 8368 int from_bitField0_ = bitField0_; 8369 int to_bitField0_ = 0; 8370 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 8371 to_bitField0_ |= 0x00000001; 8372 } 8373 if (regionBuilder_ == null) { 8374 result.region_ = region_; 8375 } else { 8376 result.region_ = regionBuilder_.build(); 8377 } 8378 result.bitField0_ = to_bitField0_; 8379 onBuilt(); 8380 return result; 8381 } 8382 mergeFrom(com.google.protobuf.Message other)8383 public Builder mergeFrom(com.google.protobuf.Message other) { 8384 if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest) { 8385 return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest)other); 8386 } else { 8387 super.mergeFrom(other); 8388 return this; 8389 } 8390 } 8391 mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest other)8392 public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest other) { 8393 if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance()) return this; 8394 if (other.hasRegion()) { 8395 mergeRegion(other.getRegion()); 8396 } 8397 this.mergeUnknownFields(other.getUnknownFields()); 8398 return this; 8399 } 8400 isInitialized()8401 public final boolean isInitialized() { 8402 if (!hasRegion()) { 8403 8404 return false; 8405 } 8406 if (!getRegion().isInitialized()) { 8407 8408 return false; 8409 } 8410 return true; 8411 } 8412 mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8413 public Builder mergeFrom( 8414 com.google.protobuf.CodedInputStream input, 8415 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8416 throws java.io.IOException { 8417 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest parsedMessage = null; 8418 try { 8419 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 8420 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8421 parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionRequest) e.getUnfinishedMessage(); 8422 throw e; 8423 } finally { 8424 if (parsedMessage != null) { 8425 mergeFrom(parsedMessage); 8426 } 8427 } 8428 return this; 8429 } 8430 private int bitField0_; 8431 8432 // required .RegionSpecifier region = 1; 8433 private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 8434 private com.google.protobuf.SingleFieldBuilder< 8435 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; 8436 /** 8437 * <code>required .RegionSpecifier region = 1;</code> 8438 */ hasRegion()8439 public boolean hasRegion() { 8440 return ((bitField0_ & 0x00000001) == 0x00000001); 8441 } 8442 /** 8443 * <code>required .RegionSpecifier region = 1;</code> 8444 */ getRegion()8445 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { 8446 if (regionBuilder_ == null) { 8447 return region_; 8448 } else { 8449 return regionBuilder_.getMessage(); 8450 } 8451 } 8452 /** 8453 * <code>required .RegionSpecifier region = 1;</code> 8454 */ setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)8455 public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 8456 if (regionBuilder_ == null) { 8457 if (value == null) { 8458 throw new NullPointerException(); 8459 } 8460 region_ = value; 8461 onChanged(); 8462 } else { 8463 regionBuilder_.setMessage(value); 8464 } 8465 bitField0_ |= 0x00000001; 8466 return this; 8467 } 8468 /** 8469 * <code>required .RegionSpecifier region = 1;</code> 8470 */ setRegion( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue)8471 public Builder setRegion( 8472 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { 8473 if (regionBuilder_ == null) { 8474 region_ = builderForValue.build(); 8475 onChanged(); 8476 } else { 8477 regionBuilder_.setMessage(builderForValue.build()); 8478 } 8479 bitField0_ |= 0x00000001; 8480 return this; 8481 } 8482 /** 8483 * <code>required .RegionSpecifier region = 1;</code> 8484 */ mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value)8485 public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) { 8486 if (regionBuilder_ == null) { 8487 if (((bitField0_ & 0x00000001) == 0x00000001) && 8488 region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { 8489 region_ = 8490 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); 8491 } else { 8492 region_ = value; 8493 } 8494 onChanged(); 8495 } else { 8496 regionBuilder_.mergeFrom(value); 8497 } 8498 bitField0_ |= 0x00000001; 8499 return this; 8500 } 8501 /** 8502 * <code>required .RegionSpecifier region = 1;</code> 8503 */ clearRegion()8504 public Builder clearRegion() { 8505 if (regionBuilder_ == null) { 8506 region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance(); 8507 onChanged(); 8508 } else { 8509 regionBuilder_.clear(); 8510 } 8511 bitField0_ = (bitField0_ & ~0x00000001); 8512 return this; 8513 } 8514 /** 8515 * <code>required .RegionSpecifier region = 1;</code> 8516 */ getRegionBuilder()8517 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { 8518 bitField0_ |= 0x00000001; 8519 onChanged(); 8520 return getRegionFieldBuilder().getBuilder(); 8521 } 8522 /** 8523 * <code>required .RegionSpecifier region = 1;</code> 8524 */ getRegionOrBuilder()8525 public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { 8526 if (regionBuilder_ != null) { 8527 return regionBuilder_.getMessageOrBuilder(); 8528 } else { 8529 return region_; 8530 } 8531 } 8532 /** 8533 * <code>required .RegionSpecifier region = 1;</code> 8534 */ 8535 private com.google.protobuf.SingleFieldBuilder< 8536 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder()8537 getRegionFieldBuilder() { 8538 if (regionBuilder_ == null) { 8539 regionBuilder_ = new com.google.protobuf.SingleFieldBuilder< 8540 org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( 8541 region_, 8542 getParentForChildren(), 8543 isClean()); 8544 region_ = null; 8545 } 8546 return regionBuilder_; 8547 } 8548 8549 // @@protoc_insertion_point(builder_scope:OfflineRegionRequest) 8550 } 8551 8552 static { 8553 defaultInstance = new OfflineRegionRequest(true); defaultInstance.initFields()8554 defaultInstance.initFields(); 8555 } 8556 8557 // @@protoc_insertion_point(class_scope:OfflineRegionRequest) 8558 } 8559 8560 public interface OfflineRegionResponseOrBuilder 8561 extends com.google.protobuf.MessageOrBuilder { 8562 } 8563 /** 8564 * Protobuf type {@code OfflineRegionResponse} 8565 */ 8566 public static final class OfflineRegionResponse extends 8567 com.google.protobuf.GeneratedMessage 8568 implements OfflineRegionResponseOrBuilder { 8569 // Use OfflineRegionResponse.newBuilder() to construct. OfflineRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder)8570 private OfflineRegionResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 8571 super(builder); 8572 this.unknownFields = builder.getUnknownFields(); 8573 } OfflineRegionResponse(boolean noInit)8574 private OfflineRegionResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 8575 8576 private static final OfflineRegionResponse defaultInstance; getDefaultInstance()8577 public static OfflineRegionResponse getDefaultInstance() { 8578 return defaultInstance; 8579 } 8580 getDefaultInstanceForType()8581 public OfflineRegionResponse getDefaultInstanceForType() { 8582 return defaultInstance; 8583 } 8584 8585 private final com.google.protobuf.UnknownFieldSet unknownFields; 8586 @java.lang.Override 8587 public final com.google.protobuf.UnknownFieldSet getUnknownFields()8588 getUnknownFields() { 8589 return this.unknownFields; 8590 } OfflineRegionResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8591 private OfflineRegionResponse( 8592 com.google.protobuf.CodedInputStream input, 8593 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8594 throws com.google.protobuf.InvalidProtocolBufferException { 8595 initFields(); 8596 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 8597 com.google.protobuf.UnknownFieldSet.newBuilder(); 8598 try { 8599 boolean done = false; 8600 while (!done) { 8601 int tag = input.readTag(); 8602 switch (tag) { 8603 case 0: 8604 done = true; 8605 break; 8606 default: { 8607 if (!parseUnknownField(input, unknownFields, 8608 extensionRegistry, tag)) { 8609 done = true; 8610 } 8611 break; 8612 } 8613 } 8614 } 8615 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 8616 throw e.setUnfinishedMessage(this); 8617 } catch (java.io.IOException e) { 8618 throw new com.google.protobuf.InvalidProtocolBufferException( 8619 e.getMessage()).setUnfinishedMessage(this); 8620 } finally { 8621 this.unknownFields = unknownFields.build(); 8622 makeExtensionsImmutable(); 8623 } 8624 } 8625 public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()8626 getDescriptor() { 8627 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionResponse_descriptor; 8628 } 8629 8630 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()8631 internalGetFieldAccessorTable() { 8632 return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_OfflineRegionResponse_fieldAccessorTable 8633 .ensureFieldAccessorsInitialized( 8634 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse.Builder.class); 8635 } 8636 8637 public static com.google.protobuf.Parser<OfflineRegionResponse> PARSER = 8638 new com.google.protobuf.AbstractParser<OfflineRegionResponse>() { 8639 public OfflineRegionResponse parsePartialFrom( 8640 com.google.protobuf.CodedInputStream input, 8641 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8642 throws com.google.protobuf.InvalidProtocolBufferException { 8643 return new OfflineRegionResponse(input, extensionRegistry); 8644 } 8645 }; 8646 8647 @java.lang.Override getParserForType()8648 public com.google.protobuf.Parser<OfflineRegionResponse> getParserForType() { 8649 return PARSER; 8650 } 8651 initFields()8652 private void initFields() { 8653 } 8654 private byte memoizedIsInitialized = -1; isInitialized()8655 public final boolean isInitialized() { 8656 byte isInitialized = memoizedIsInitialized; 8657 if (isInitialized != -1) return isInitialized == 1; 8658 8659 memoizedIsInitialized = 1; 8660 return true; 8661 } 8662 writeTo(com.google.protobuf.CodedOutputStream output)8663 public void writeTo(com.google.protobuf.CodedOutputStream output) 8664 throws java.io.IOException { 8665 getSerializedSize(); 8666 getUnknownFields().writeTo(output); 8667 } 8668 8669 private int memoizedSerializedSize = -1; getSerializedSize()8670 public int getSerializedSize() { 8671 int size = memoizedSerializedSize; 8672 if (size != -1) return size; 8673 8674 size = 0; 8675 size += getUnknownFields().getSerializedSize(); 8676 memoizedSerializedSize = size; 8677 return size; 8678 } 8679 8680 private static final long serialVersionUID = 0L; 8681 @java.lang.Override writeReplace()8682 protected java.lang.Object writeReplace() 8683 throws java.io.ObjectStreamException { 8684 return super.writeReplace(); 8685 } 8686 8687 @java.lang.Override equals(final java.lang.Object obj)8688 public boolean equals(final java.lang.Object obj) { 8689 if (obj == this) { 8690 return true; 8691 } 8692 if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse)) { 8693 return super.equals(obj); 8694 } 8695 org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse) obj; 8696 8697 boolean result = true; 8698 result = result && 8699 getUnknownFields().equals(other.getUnknownFields()); 8700 return result; 8701 } 8702 8703 private int memoizedHashCode = 0; 8704 @java.lang.Override hashCode()8705 public int hashCode() { 8706 if (memoizedHashCode != 0) { 8707 return memoizedHashCode; 8708 } 8709 int hash = 41; 8710 hash = (19 * hash) + getDescriptorForType().hashCode(); 8711 hash = (29 * hash) + getUnknownFields().hashCode(); 8712 memoizedHashCode = hash; 8713 return hash; 8714 } 8715 parseFrom( com.google.protobuf.ByteString data)8716 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( 8717 com.google.protobuf.ByteString data) 8718 throws com.google.protobuf.InvalidProtocolBufferException { 8719 return PARSER.parseFrom(data); 8720 } parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8721 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( 8722 com.google.protobuf.ByteString data, 8723 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8724 throws com.google.protobuf.InvalidProtocolBufferException { 8725 return PARSER.parseFrom(data, extensionRegistry); 8726 } parseFrom(byte[] data)8727 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom(byte[] data) 8728 throws com.google.protobuf.InvalidProtocolBufferException { 8729 return PARSER.parseFrom(data); 8730 } parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8731 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( 8732 byte[] data, 8733 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8734 throws com.google.protobuf.InvalidProtocolBufferException { 8735 return PARSER.parseFrom(data, extensionRegistry); 8736 } parseFrom(java.io.InputStream input)8737 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom(java.io.InputStream input) 8738 throws java.io.IOException { 8739 return PARSER.parseFrom(input); 8740 } parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)8741 public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( 8742 java.io.InputStream input, 8743 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 8744 throws java.io.IOException { 8745 return PARSER.parseFrom(input, extensionRegistry); 8746 } parseDelimitedFrom(java.io.InputStream input)8747