if (subBuilder != null) { subBuilder.mergeFrom(readOpChecksumInfo_); readOpChecksumInfo_ = subBuilder.buildPartial();
if (subBuilder != null) { subBuilder.mergeFrom(readOpChecksumInfo_); readOpChecksumInfo_ = subBuilder.buildPartial();
/** * <code>optional .hadoop.hdfs.ReadOpChecksumInfoProto readOpChecksumInfo = 4;</code> */ public Builder mergeReadOpChecksumInfo(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto value) { if (readOpChecksumInfoBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && readOpChecksumInfo_ != org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.getDefaultInstance()) { readOpChecksumInfo_ = org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.newBuilder(readOpChecksumInfo_).mergeFrom(value).buildPartial(); } else { readOpChecksumInfo_ = value; } onChanged(); } else { readOpChecksumInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /**
if (subBuilder != null) { subBuilder.mergeFrom(readOpChecksumInfo_); readOpChecksumInfo_ = subBuilder.buildPartial();
/** * <code>optional .hadoop.hdfs.ReadOpChecksumInfoProto readOpChecksumInfo = 4;</code> */ public Builder mergeReadOpChecksumInfo(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto value) { if (readOpChecksumInfoBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && readOpChecksumInfo_ != org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.getDefaultInstance()) { readOpChecksumInfo_ = org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.newBuilder(readOpChecksumInfo_).mergeFrom(value).buildPartial(); } else { readOpChecksumInfo_ = value; } onChanged(); } else { readOpChecksumInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /**
/** * <code>optional .hadoop.hdfs.ReadOpChecksumInfoProto readOpChecksumInfo = 4;</code> */ public Builder mergeReadOpChecksumInfo(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto value) { if (readOpChecksumInfoBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && readOpChecksumInfo_ != org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.getDefaultInstance()) { readOpChecksumInfo_ = org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto.newBuilder(readOpChecksumInfo_).mergeFrom(value).buildPartial(); } else { readOpChecksumInfo_ = value; } onChanged(); } else { readOpChecksumInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /**
public Builder clone() { return create().mergeFrom(buildPartial()); }
public Builder clone() { return create().mergeFrom(buildPartial()); }
public Builder clone() { return create().mergeFrom(buildPartial()); }
public org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto build() { org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }
public org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto build() { org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }
public org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto build() { org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.ReadOpChecksumInfoProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; }