/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /**
finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } else { finalizeCmdBuilder_.clear();
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } else { finalizeCmdBuilder_.clear();
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /**
public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); }
public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); }
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /**
/** * <code>optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5;</code> */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /**
/** * <code>required string blockPoolId = 1;</code> * * <pre> * Block pool to be finalized * </pre> */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /**
/** * <code>required string blockPoolId = 1;</code> * * <pre> * Block pool to be finalized * </pre> */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /**
private void initFields() { cmdType_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.BalancerBandwidthCommand; balancerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance(); blkCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance(); recoveryCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance(); finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); keyUpdateCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance(); registerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance(); blkIdCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
private void initFields() { cmdType_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.BalancerBandwidthCommand; balancerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance(); blkCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance(); recoveryCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance(); finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); keyUpdateCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance(); registerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance(); blkIdCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) return this; if (other.hasBlockPoolId()) { bitField0_ |= 0x00000001; blockPoolId_ = other.blockPoolId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) return this; if (other.hasBlockPoolId()) { bitField0_ |= 0x00000001; blockPoolId_ = other.blockPoolId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
/** * <code>required string blockPoolId = 1;</code> * * <pre> * Block pool to be finalized * </pre> */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /**
public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); }
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) return this; if (other.hasBlockPoolId()) { bitField0_ |= 0x00000001; blockPoolId_ = other.blockPoolId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; }
private void initFields() { cmdType_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.BalancerBandwidthCommand; balancerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance(); blkCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance(); recoveryCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance(); finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); keyUpdateCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance(); registerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance(); blkIdCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance(); blkECReconstructionCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1;