public class ParquetRecordReader extends AbstractRecordReader
| Modifier and Type | Class and Description |
|---|---|
static class |
ParquetRecordReader.Metric |
| Modifier and Type | Field and Description |
|---|---|
int |
bufferedReadSize |
protected static List<SchemaPath> |
DEFAULT_COLS_TO_READ |
static byte[] |
endBitMasks |
boolean |
enforceTotalSize |
static int |
PARQUET_PAGE_MAX_SIZE |
ParquetReaderStats |
parquetReaderStats |
long |
readQueueSize |
static byte[] |
startBitMasks |
boolean |
useAsyncColReader |
boolean |
useAsyncPageReader |
boolean |
useBufferedReader |
boolean |
useFadvise |
DEFAULT_TEXT_COLS_TO_READALLOCATOR_INITIAL_RESERVATION, ALLOCATOR_MAX_RESERVATION| Constructor and Description |
|---|
ParquetRecordReader(FragmentContext fragmentContext,
long batchSize,
long numRecordsToRead,
String path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
String path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
String path,
int rowGroupIndex,
long numRecordsToRead,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.hadoop.CodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
| Modifier and Type | Method and Description |
|---|---|
void |
allocate(Map<String,ValueVector> vectorMap) |
void |
close() |
long |
getBatchSize() |
int |
getBitWidthAllFixedFields() |
org.apache.parquet.hadoop.CodecFactory |
getCodecFactory() |
ParquetReaderUtility.DateCorruptionStatus |
getDateCorruptionStatus()
Flag indicating if the old non-standard data format appears
in this file, see DRILL-4203.
|
protected List<SchemaPath> |
getDefaultColumnsToRead() |
org.apache.hadoop.fs.FileSystem |
getFileSystem() |
FragmentContext |
getFragmentContext() |
org.apache.hadoop.fs.Path |
getHadoopPath() |
OperatorContext |
getOperatorContext() |
int |
getRowGroupIndex() |
protected DrillRuntimeException |
handleException(String s,
Exception e) |
int |
next()
Read the next record batch from the file using the reader and read state
created previously.
|
void |
setup(OperatorContext operatorContext,
OutputMutator output)
Prepare the Parquet reader.
|
getColumns, isSkipQuery, isStarQuery, setColumns, toString, transformColumnsprotected static final List<SchemaPath> DEFAULT_COLS_TO_READ
public static final int PARQUET_PAGE_MAX_SIZE
public static final byte[] endBitMasks
public static final byte[] startBitMasks
public boolean useAsyncColReader
public boolean useAsyncPageReader
public boolean useBufferedReader
public int bufferedReadSize
public boolean useFadvise
public boolean enforceTotalSize
public long readQueueSize
public ParquetReaderStats parquetReaderStats
public ParquetRecordReader(FragmentContext fragmentContext, String path, int rowGroupIndex, long numRecordsToRead, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) throws ExecutionSetupException
ExecutionSetupExceptionpublic ParquetRecordReader(FragmentContext fragmentContext, String path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) throws ExecutionSetupException
ExecutionSetupExceptionpublic ParquetRecordReader(FragmentContext fragmentContext, long batchSize, long numRecordsToRead, String path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.hadoop.CodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) throws ExecutionSetupException
ExecutionSetupExceptionpublic ParquetReaderUtility.DateCorruptionStatus getDateCorruptionStatus()
public org.apache.parquet.hadoop.CodecFactory getCodecFactory()
public org.apache.hadoop.fs.Path getHadoopPath()
public org.apache.hadoop.fs.FileSystem getFileSystem()
public int getRowGroupIndex()
public int getBitWidthAllFixedFields()
public long getBatchSize()
public OperatorContext getOperatorContext()
public FragmentContext getFragmentContext()
public void setup(OperatorContext operatorContext, OutputMutator output) throws ExecutionSetupException
operatorContext - operator context for the readeroutput - The place where output for a particular scan should be written. The record reader is responsible for
mutating the set of schema values for that particular record.ExecutionSetupExceptionprotected DrillRuntimeException handleException(String s, Exception e)
public void allocate(Map<String,ValueVector> vectorMap) throws OutOfMemoryException
allocate in interface RecordReaderallocate in class AbstractRecordReaderOutOfMemoryExceptionpublic int next()
public void close()
protected List<SchemaPath> getDefaultColumnsToRead()
getDefaultColumnsToRead in class AbstractRecordReaderCopyright © 2017 The Apache Software Foundation. All rights reserved.