public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); recoveryId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); newLength_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); newBlockId_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; }
public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); newGenStamp_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); newLength_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); closeFile_ = false; bitField0_ = (bitField0_ & ~0x00000008); deleteBlock_ = false; bitField0_ = (bitField0_ & ~0x00000010); if (newTaragetsBuilder_ == null) { newTaragets_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); } else { newTaragetsBuilder_.clear(); } newTargetStorages_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000040); return this; }
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); newGenStamp_ = 0L; newLength_ = 0L; closeFile_ = false; deleteBlock_ = false; newTaragets_ = java.util.Collections.emptyList(); newTargetStorages_ = com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); recoveryId_ = 0L; newLength_ = 0L; newBlockId_ = 0L; } private byte memoizedIsInitialized = -1;
public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; }
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); localPath_ = ""; localMetaPath_ = ""; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); token_ = org.apache.hadoop.security.proto.SecurityProtos.TokenProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); newGenStamp_ = 0L; newLength_ = 0L; closeFile_ = false; deleteBlock_ = false; newTaragets_ = java.util.Collections.emptyList(); newTargetStorages_ = com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); localPath_ = ""; localMetaPath_ = ""; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); token_ = org.apache.hadoop.security.proto.SecurityProtos.TokenProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; }
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); newGenStamp_ = 0L; newLength_ = 0L; closeFile_ = false; deleteBlock_ = false; newTaragets_ = java.util.Collections.emptyList(); newTargetStorages_ = io.prestosql.hadoop.$internal.com.google.protobuf.LazyStringArrayList.EMPTY; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); recoveryId_ = 0L; newLength_ = 0L; newBlockId_ = 0L; } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); token_ = org.apache.hadoop.security.proto.SecurityProtos.TokenProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); sourceDnInfos_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.getDefaultInstance(); targetDnInfos_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.getDefaultInstance(); targetStorageUuids_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.getDefaultInstance(); targetStorageTypes_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.getDefaultInstance(); liveBlockIndices_ = com.google.protobuf.ByteString.EMPTY; ecPolicy_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ErasureCodingPolicyProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
/** * <code>required .hadoop.hdfs.ExtendedBlockProto block = 1;</code> * * <pre> * Block identifier * </pre> */ public Builder mergeBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto value) { if (blockBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && block_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance()) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.newBuilder(block_).mergeFrom(value).buildPartial(); } else { block_ = value; } onChanged(); } else { blockBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /**
/** * <code>required .hadoop.hdfs.ExtendedBlockProto block = 1;</code> */ public Builder mergeBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto value) { if (blockBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && block_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance()) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.newBuilder(block_).mergeFrom(value).buildPartial(); } else { block_ = value; } onChanged(); } else { blockBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /**
/** * <code>required .hadoop.hdfs.ExtendedBlockProto block = 1;</code> * * <pre> * Block identifier * </pre> */ public Builder clearBlock() { if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); onChanged(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /**
/** * <code>required .hadoop.hdfs.ExtendedBlockProto block = 1;</code> */ public Builder clearBlock() { if (blockBuilder_ == null) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); onChanged(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /**
private void initFields() { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance(); recoveryId_ = 0L; newLength_ = 0L; newBlockId_ = 0L; } private byte memoizedIsInitialized = -1;