.compress(reader.getCompression()) .version(reader.getFileVersion()) .rowIndexStride(reader.getRowIndexStride()) .inspector(reader.getObjectInspector());
.compress(CompressionKind.NONE) .bufferSize(100) .rowIndexStride(1000)); Random r1 = new Random(1); Random r2 = new Random(2);
.compress(CompressionKind.SNAPPY) .bufferSize(1000) .rowIndexStride(0)); Random rand = new Random(24); for(int i=0; i < 10000; ++i) {
.blockPadding(false) .compress(CompressionKind.NONE) .rowIndexStride(0);
.compress(compress) .bufferSize(bufferSize) .rowIndexStride(rowIndexStride));
static WriterOptions createOrcWriterOptions(ObjectInspector sourceOi, Configuration conf, CacheWriter cacheWriter, int allocSize) throws IOException { return OrcFile.writerOptions(conf).stripeSize(Long.MAX_VALUE).blockSize(Long.MAX_VALUE) .rowIndexStride(Integer.MAX_VALUE) // For now, do not limit this - one RG per split .blockPadding(false).compress(CompressionKind.NONE).version(Version.CURRENT) .encodingStrategy(EncodingStrategy.SPEED).bloomFilterColumns(null).inspector(sourceOi) .physicalWriter(cacheWriter).memory(MEMORY_MANAGER).bufferSize(allocSize); }
.compress(compression) .version(version) .rowIndexStride(rowIndexStride) .inspector(reader.getObjectInspector());
.stripeSize(200000) .bufferSize(65536) .rowIndexStride(1000)); Random rand = new Random(42); final int COUNT=32768;
HiveConf.ConfVars.HIVE_ORC_DELTA_STREAMING_OPTIMIZATIONS_ENABLED.varname, false)) { writerOptions.encodingStrategy(org.apache.orc.OrcFile.EncodingStrategy.SPEED); writerOptions.rowIndexStride(0); writerOptions.getConfiguration().set(OrcConf.DICTIONARY_KEY_SIZE_THRESHOLD.getAttribute(), "-1.0");
.stripeSize(200000) .bufferSize(65536) .rowIndexStride(1000)); Random rand = new Random(42); final int COUNT=32768;
.compress(compression) .version(fileVersion) .rowIndexStride(rowIndexStride) .inspector(reader.getObjectInspector());
.rowIndexStride(rowIndexStride) .blockPadding(addBlockPadding) .version(versionValue)
.blockPadding(false) .compress(CompressionKind.NONE) .rowIndexStride(0);
if(!MetastoreConf.getBoolVar(options.getConfiguration(), MetastoreConf.ConfVars.COMPACTOR_MINOR_STATS_COMPRESSION)) { opts.compress(CompressionKind.NONE).rowIndexStride(0);
.stripeSize(50000) .bufferSize(100) .rowIndexStride(0) .memory(memory) .batchSize(100)
.stripeSize(50000) .bufferSize(100) .rowIndexStride(0) .memory(memory) .batchSize(100)
.compress(compress) .bufferSize(bufferSize) .rowIndexStride(rowIndexStride));
.compress(compress) .bufferSize(bufferSize) .rowIndexStride(rowIndexStride));
private void createOrcFile(int stripSize, int numberOfRows) throws IOException { ObjectInspector inspector; synchronized (TestOrcFileStripeMergeRecordReader.class) { inspector = ObjectInspectorFactory.getReflectionObjectInspector (StringIntIntIntRow.class, ObjectInspectorFactory.ObjectInspectorOptions.JAVA); } Writer writer = OrcFile.createWriter(tmpPath, OrcFile.writerOptions(conf) .inspector(inspector) .stripeSize(stripSize) .compress(CompressionKind.ZLIB) .bufferSize(5000) .rowIndexStride(1000)); Random rand = new Random(157); for (int i = 0; i < numberOfRows; i++) { writer.addRow(new StringIntIntIntRow( Integer.toBinaryString(i), rand.nextInt(), rand.nextInt(), rand.nextInt() )); } writer.close(); }
static WriterOptions createOrcWriterOptions(ObjectInspector sourceOi, Configuration conf, CacheWriter cacheWriter, int allocSize) throws IOException { return OrcFile.writerOptions(conf).stripeSize(Long.MAX_VALUE).blockSize(Long.MAX_VALUE) .rowIndexStride(Integer.MAX_VALUE) // For now, do not limit this - one RG per split .blockPadding(false).compress(CompressionKind.NONE).version(Version.CURRENT) .encodingStrategy(EncodingStrategy.SPEED).bloomFilterColumns(null).inspector(sourceOi) .physicalWriter(cacheWriter).memory(MEMORY_MANAGER).bufferSize(allocSize); }