ChecksumProto checksumProto = DataTransferProtoUtil.toProto(summer); OpWriteBlockProto.Builder writeBlockProtoBuilder = OpWriteBlockProto.newBuilder() .setHeader(header).setStage(OpWriteBlockProto.BlockConstructionStage.valueOf(stage.name())) .setPipelineSize(1).setMinBytesRcvd(locatedBlock.getBlock().getNumBytes()) .setMaxBytesRcvd(maxBytesRcvd).setLatestGenerationStamp(latestGS) .setRequestedChecksum(checksumProto) .setCachingStrategy(CachingStrategyProto.newBuilder().setDropBehind(true).build()); List<Future<Channel>> futureList = new ArrayList<>(datanodeInfos.length); for (int i = 0; i < datanodeInfos.length; i++) {
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto.getDefaultInstance()) return this; if (other.hasHeader()) { mergeHeader(other.getHeader()); bitField0_ = (bitField0_ & ~0x00000002); } else { ensureTargetsIsMutable(); targets_.addAll(other.targets_); onChanged(); targetsBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getTargetsFieldBuilder() : null; } else { targetsBuilder_.addAllMessages(other.targets_); mergeSource(other.getSource()); setStage(other.getStage()); setPipelineSize(other.getPipelineSize()); setMinBytesRcvd(other.getMinBytesRcvd()); setMaxBytesRcvd(other.getMaxBytesRcvd()); setLatestGenerationStamp(other.getLatestGenerationStamp());
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto.getDefaultInstance()) return this; if (other.hasHeader()) { mergeHeader(other.getHeader()); bitField0_ = (bitField0_ & ~0x00000002); } else { ensureTargetsIsMutable(); targets_.addAll(other.targets_); onChanged(); targetsBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getTargetsFieldBuilder() : null; } else { targetsBuilder_.addAllMessages(other.targets_); mergeSource(other.getSource()); setStage(other.getStage()); setPipelineSize(other.getPipelineSize()); setMinBytesRcvd(other.getMinBytesRcvd()); setMaxBytesRcvd(other.getMaxBytesRcvd()); setLatestGenerationStamp(other.getLatestGenerationStamp());
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.OpWriteBlockProto.getDefaultInstance()) return this; if (other.hasHeader()) { mergeHeader(other.getHeader()); bitField0_ = (bitField0_ & ~0x00000002); } else { ensureTargetsIsMutable(); targets_.addAll(other.targets_); onChanged(); targetsBuilder_ = io.prestosql.hadoop.$internal.com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getTargetsFieldBuilder() : null; } else { targetsBuilder_.addAllMessages(other.targets_); mergeSource(other.getSource()); setStage(other.getStage()); setPipelineSize(other.getPipelineSize()); setMinBytesRcvd(other.getMinBytesRcvd()); setMaxBytesRcvd(other.getMaxBytesRcvd()); setLatestGenerationStamp(other.getLatestGenerationStamp());
.setHeader(header) .setStorageType(PBHelperClient.convertStorageType(storageType)) .addAllTargets(PBHelperClient.convert(targets, 1)) .addAllTargetStorageTypes( PBHelperClient.convertStorageTypes(targetStorageTypes, 1)) .setStage(toProto(stage)) .setPipelineSize(pipelineSize) .setMinBytesRcvd(minBytesRcvd) .setMaxBytesRcvd(maxBytesRcvd) .setLatestGenerationStamp(latestGenerationStamp) .setRequestedChecksum(checksumProto) .setCachingStrategy(getCachingStrategy(cachingStrategy)) .setAllowLazyPersist(allowLazyPersist) .setPinning(pinning) .addAllTargetPinnings(PBHelperClient.convert(targetPinnings, 1)) .addAllTargetStorageIds(PBHelperClient.convert(targetStorageIds, 1)); if (source != null) { proto.setSource(PBHelperClient.convertDatanodeInfo(source)); proto.setStorageId(storageId); send(out, Op.WRITE_BLOCK, proto.build());
.setHeader(header) .setStorageType(PBHelper.convertStorageType(storageType)) .addAllTargets(PBHelper.convert(targets, 1)) .addAllTargetStorageTypes(PBHelper.convertStorageTypes(targetStorageTypes, 1)) .setStage(toProto(stage)) .setPipelineSize(pipelineSize) .setMinBytesRcvd(minBytesRcvd) .setMaxBytesRcvd(maxBytesRcvd) .setLatestGenerationStamp(latestGenerationStamp) .setRequestedChecksum(checksumProto) .setCachingStrategy(getCachingStrategy(cachingStrategy)) .setAllowLazyPersist(allowLazyPersist) .setPinning(pinning) .addAllTargetPinnings(PBHelper.convert(targetPinnings, 1)); proto.setSource(PBHelper.convertDatanodeInfo(source)); send(out, Op.WRITE_BLOCK, proto.build());
.setHeader(header) .setStorageType(PBHelper.convertStorageType(storageType)) .addAllTargets(PBHelper.convert(targets, 1)) .addAllTargetStorageTypes(PBHelper.convertStorageTypes(targetStorageTypes, 1)) .setStage(toProto(stage)) .setPipelineSize(pipelineSize) .setMinBytesRcvd(minBytesRcvd) .setMaxBytesRcvd(maxBytesRcvd) .setLatestGenerationStamp(latestGenerationStamp) .setRequestedChecksum(checksumProto) .setCachingStrategy(getCachingStrategy(cachingStrategy)) .setAllowLazyPersist(allowLazyPersist) .setPinning(pinning) .addAllTargetPinnings(PBHelper.convert(targetPinnings, 1)); proto.setSource(PBHelper.convertDatanodeInfo(source)); send(out, Op.WRITE_BLOCK, proto.build());
private static Builder create() { return new Builder(); }
private static Builder create() { return new Builder(); }
private static Builder create() { return new Builder(); }
@java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /**
@java.lang.Override protected Builder newBuilderForType( io.prestosql.hadoop.$internal.com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /**
@java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /**
private static void requestWriteBlock(Channel channel, Enum<?> storageType, OpWriteBlockProto.Builder writeBlockProtoBuilder) throws IOException { OpWriteBlockProto proto = STORAGE_TYPE_SETTER.set(writeBlockProtoBuilder, storageType).build(); int protoLen = proto.getSerializedSize(); ByteBuf buffer = channel.alloc().buffer(3 + CodedOutputStream.computeRawVarint32Size(protoLen) + protoLen); buffer.writeShort(DataTransferProtocol.DATA_TRANSFER_VERSION); buffer.writeByte(Op.WRITE_BLOCK.code); proto.writeDelimitedTo(new ByteBufOutputStream(buffer)); channel.writeAndFlush(buffer); }