org.apache.hadoop.mapreduce.RecordReader<org.apache.hadoop.io.NullWritable,OrcStruct> |
OrcNewInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
AcidInputFormat.RawReader<OrcStruct> |
OrcInputFormat.getRawReader(org.apache.hadoop.conf.Configuration conf,
boolean collapseEvents,
int bucket,
ValidTxnList validTxnList,
org.apache.hadoop.fs.Path baseDirectory,
org.apache.hadoop.fs.Path[] deltaDirectory) |
AcidInputFormat.RowReader<OrcStruct> |
OrcInputFormat.getReader(org.apache.hadoop.mapred.InputSplit inputSplit,
AcidInputFormat.Options options) |
org.apache.hadoop.mapred.RecordReader<org.apache.hadoop.io.NullWritable,OrcStruct> |
OrcInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit inputSplit,
org.apache.hadoop.mapred.JobConf conf,
org.apache.hadoop.mapred.Reporter reporter) |