Interface | Description |
---|---|
AcidInputFormat<KEY extends org.apache.hadoop.io.WritableComparable,VALUE> |
The interface required for input formats that what to support ACID
transactions.
|
AcidInputFormat.AcidRecordReader<K,V> |
RecordReader returned by AcidInputFormat working in row-at-a-time mode should AcidRecordReader.
|
AcidInputFormat.RawReader<V> | |
AcidInputFormat.RowReader<V> | |
AcidOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V> |
An extension for OutputFormats that want to implement ACID transactions.
|
AcidUtils.Directory | |
CombineHiveInputFormat.AvoidSplitCombination |
This is a marker interface that is used to identify the formats where
combine split generation is not applicable
|
ContentSummaryInputFormat |
ContentSummayInputFormat provides an interface to let the input format itself
figure the content summary for a give input path.
|
FlatFileInputFormat.SerializationContext<S> |
An implementation of SerializationContext is responsible for looking up the
Serialization implementation for the given RecordReader.
|
HiveOutputFormat<K,V> |
HiveOutputFormat describes the output-specification for Hive's
operators. |
HivePartitioner<K2,V2> |
Partition keys by their
Object.hashCode() . |
InputFormatChecker |
Check for validity of the input files.
|
RecordUpdater |
API for supporting updating records.
|
ReworkMapredInputFormat | |
StatsProvidingRecordReader |
If a file format internally gathers statistics (like ORC) then it can expose
the statistics through this interface.
|
StatsProvidingRecordWriter |
If a file format internally gathers statistics (like ORC) while writing then
it can expose the statistics through this record writer interface.
|
StorageFormatDescriptor |
Subclasses represent a storage format for the
CREATE TABLE ...
|
Class | Description |
---|---|
AbstractStorageFormatDescriptor | |
AcidInputFormat.Options |
Options for controlling the record readers.
|
AcidOutputFormat.Options |
Options to control how the files are written
|
AcidUtils |
Utilities that are shared by all of the ACID input and output formats.
|
AcidUtils.ParsedDelta | |
AvroStorageFormatDescriptor | |
BucketizedHiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
BucketizedHiveInputFormat serves the similar function as hiveInputFormat but
its getSplits() always group splits from one input file into one wrapper
split.
|
BucketizedHiveInputSplit |
HiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClass.
|
BucketizedHiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
BucketizedHiveRecordReader is a wrapper on a list of RecordReader.
|
CodecPool |
A global compressor/decompressor pool used to save and reuse (possibly
native) compression/decompression codecs.
|
CombineHiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
CombineHiveInputFormat is a parameterized InputFormat which looks at the path
name and determine the correct InputFormat for that path name from
mapredPlan.pathToPartitionInfo().
|
CombineHiveInputFormat.CombineHiveInputSplit |
CombineHiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClassName.
|
CombineHiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
CombineHiveRecordReader.
|
DefaultHivePartitioner<K2,V2> |
Partition keys by their
Object.hashCode() . |
FlatFileInputFormat<T> | Deprecated |
FlatFileInputFormat.RowContainer<T> |
A work-around until HADOOP-1230 is fixed.
|
FlatFileInputFormat.SerializationContextFromConf<S> |
An implementation of
FlatFileInputFormat.SerializationContext that reads the
Serialization class and specific subclass to be deserialized from the
JobConf. |
HiveBinaryOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveBinaryOutputFormat writes out the values consecutively without any
separators.
|
HiveContextAwareRecordReader<K,V> |
This class prepares an IOContext, and provides the ability to perform a binary search on the
data.
|
HiveFileFormatUtils |
An util class for various Hive file format tasks.
|
HiveFileFormatUtils.NullOutputCommitter | |
HiveIgnoreKeyTextOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveIgnoreKeyTextOutputFormat replaces key with null before feeding the
|
HiveIgnoreKeyTextOutputFormat.IgnoreKeyWriter<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | |
HiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveInputFormat is a parameterized InputFormat which looks at the path name
and determine the correct InputFormat for that path name from
mapredPlan.pathToPartitionInfo().
|
HiveInputFormat.HiveInputSplit |
HiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClass.
|
HiveKey |
HiveKey is a simple wrapper on Text which allows us to set the hashCode
easily.
|
HiveKey.Comparator |
A Comparator optimized for HiveKey.
|
HiveNullValueSequenceFileOutputFormat<K,V> |
A
HiveOutputFormat that writes SequenceFile s with the
content saved in the keys, and null in the values. |
HiveOutputFormatImpl<K extends org.apache.hadoop.io.WritableComparable<K>,V extends org.apache.hadoop.io.Writable> |
Hive does not use OutputFormat's in a conventional way, but constructs and uses
the defined OutputFormat for each table from FileSinkOperator.
|
HivePassThroughOutputFormat<K,V> |
This pass through class is used to wrap OutputFormat implementations such that new OutputFormats not derived from
HiveOutputFormat gets through the checker
|
HivePassThroughRecordWriter<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> | |
HiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveRecordReader is a simple wrapper on RecordReader.
|
HiveSequenceFileOutputFormat<K,V> |
A
HiveOutputFormat that writes SequenceFile s. |
IgnoreKeyTextOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | Deprecated
use
HiveIgnoreKeyTextOutputFormat instead} |
IgnoreKeyTextOutputFormat.IgnoreKeyWriter<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | |
IOConstants | |
IOContext |
IOContext basically contains the position information of the current
key/value.
|
IOPrepareCache |
IOPrepareCache is used to cache pre-query io-related objects.
|
NonSyncDataInputBuffer |
A thread-not-safe version of Hadoop's DataInputBuffer, which removes all
synchronized modifiers.
|
NonSyncDataOutputBuffer |
A thread-not-safe version of Hadoop's DataOutputBuffer, which removes all
synchronized modifiers.
|
NullRowsInputFormat |
NullRowsInputFormat outputs null rows, maximum 100.
|
NullRowsInputFormat.DummyInputSplit | |
NullRowsInputFormat.NullRowsRecordReader | |
OneNullRowInputFormat |
OneNullRowInputFormat outputs one null row.
|
OneNullRowInputFormat.OneNullRowRecordReader | |
ORCFileStorageFormatDescriptor | |
ParquetFileStorageFormatDescriptor | |
RCFile |
RCFile s, short of Record Columnar File, are flat files
consisting of binary key/value pairs, which shares much similarity with
SequenceFile . |
RCFile.KeyBuffer |
KeyBuffer is the key of each record in RCFile.
|
RCFile.Reader |
Read KeyBuffer/ValueBuffer pairs from a RCFile.
|
RCFile.ValueBuffer |
ValueBuffer is the value of each record in RCFile.
|
RCFile.Writer |
Write KeyBuffer/ValueBuffer pairs to a RCFile.
|
RCFileInputFormat<K extends org.apache.hadoop.io.LongWritable,V extends BytesRefArrayWritable> |
RCFileInputFormat.
|
RCFileOutputFormat |
RCFileOutputFormat.
|
RCFileRecordReader<K extends org.apache.hadoop.io.LongWritable,V extends BytesRefArrayWritable> |
RCFileRecordReader.
|
RCFileStorageFormatDescriptor | |
RecordIdentifier |
Gives the Record identifier information for the current record.
|
RecordIdentifier.StructInfo |
RecordIdentifier is passed along the operator tree as a struct.
|
SchemaAwareCompressionInputStream |
SchemaAwareCompressionInputStream adds the ability to inform the compression
stream what column is being read.
|
SchemaAwareCompressionOutputStream |
SchemaAwareCompressionOutputStream adds the ability to inform the comression stream
the current column being compressed.
|
SequenceFileInputFormatChecker |
SequenceFileInputFormatChecker.
|
SequenceFileStorageFormatDescriptor | |
StorageFormatFactory | |
SymbolicInputFormat | |
SymlinkTextInputFormat |
Symlink file is a text file which contains a list of filename / dirname.
|
SymlinkTextInputFormat.SymlinkTextInputSplit |
This input split wraps the FileSplit generated from
TextInputFormat.getSplits(), while setting the original link file path
as job input path.
|
TextFileStorageFormatDescriptor | |
VectorizedRCFileInputFormat |
A MapReduce/Hive Vectorized input format for RC files.
|
VectorizedRCFileRecordReader |
RCFileRecordReader.
|
Enum | Description |
---|---|
AcidUtils.Operation | |
IOContext.Comparison | |
RecordIdentifier.Field |
This is in support of
VirtualColumn.ROWID
Contains metadata about each field in RecordIdentifier that needs to be part of ROWID
which is represented as a struct RecordIdentifier.StructInfo . |
Copyright © 2017 The Apache Software Foundation. All rights reserved.