Modifier and Type | Method and Description |
---|---|
protected LazyObjectBase |
LazyAccumuloRow.createLazyField(int fieldID,
StructField fieldRef) |
Modifier and Type | Method and Description |
---|---|
protected void |
HiveAccumuloTableInputFormat.configure(org.apache.hadoop.mapred.JobConf conf,
org.apache.accumulo.core.client.Instance instance,
org.apache.accumulo.core.client.Connector connector,
AccumuloConnectionParameters accumuloParams,
ColumnMapper columnMapper,
List<org.apache.accumulo.core.client.IteratorSetting> iterators,
Collection<org.apache.accumulo.core.data.Range> ranges)
Configure the underlying AccumuloInputFormat
|
Modifier and Type | Method and Description |
---|---|
CompareOp |
AccumuloPredicateHandler.getCompareOp(String udfType,
IndexSearchCondition sc) |
byte[] |
PushdownTuple.getConstantAsBytes(org.apache.hadoop.io.Writable writable) |
List<org.apache.accumulo.core.client.IteratorSetting> |
AccumuloPredicateHandler.getIterators(org.apache.hadoop.conf.Configuration conf,
ColumnMapper columnMapper)
Loop through search conditions and build iterator settings for predicates involving columns
other than rowID, if any.
|
PrimitiveComparison |
AccumuloPredicateHandler.getPrimitiveComparison(String type,
IndexSearchCondition sc) |
List<org.apache.accumulo.core.data.Range> |
AccumuloPredicateHandler.getRanges(org.apache.hadoop.conf.Configuration conf,
ColumnMapper columnMapper)
Loop through search conditions and build ranges for predicates involving rowID column, if any.
|
org.apache.accumulo.core.client.IteratorSetting |
AccumuloPredicateHandler.toSetting(HiveAccumuloColumnMapping accumuloColumnMapping,
IndexSearchCondition sc)
Create an IteratorSetting for the right qualifier, constant, CompareOpt, and PrimitiveCompare
type.
|
Constructor and Description |
---|
PushdownTuple(IndexSearchCondition sc,
PrimitiveComparison pCompare,
CompareOp cOpt) |
Modifier and Type | Class and Description |
---|---|
class |
TooManyAccumuloColumnsException |
class |
TooManyHiveColumnsException |
Modifier and Type | Method and Description |
---|---|
LazyObjectBase |
DefaultAccumuloRowIdFactory.createRowId(ObjectInspector inspector) |
T |
CompositeAccumuloRowIdFactory.createRowId(ObjectInspector inspector) |
LazyObjectBase |
AccumuloRowIdFactory.createRowId(ObjectInspector inspector)
create custom object for accumulo
|
ObjectInspector |
DefaultAccumuloRowIdFactory.createRowIdObjectInspector(TypeInfo type) |
ObjectInspector |
AccumuloRowIdFactory.createRowIdObjectInspector(TypeInfo type)
create custom object inspector for accumulo rowId
|
Object |
AccumuloSerDe.deserialize(org.apache.hadoop.io.Writable writable) |
protected ArrayList<ObjectInspector> |
AccumuloSerDe.getColumnObjectInspectors(List<TypeInfo> columnTypes,
LazySerDeParameters serDeParams,
List<ColumnMapping> mappings,
AccumuloRowIdFactory factory) |
ObjectInspector |
AccumuloSerDe.getObjectInspector() |
void |
DefaultAccumuloRowIdFactory.init(AccumuloSerDeParameters accumuloSerDeParams,
Properties properties) |
void |
AccumuloRowIdFactory.init(AccumuloSerDeParameters serDeParams,
Properties properties)
initialize factory with properties
|
void |
AccumuloSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties properties) |
protected AccumuloRowIdFactory |
AccumuloSerDeParameters.initRowIdFactory(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
org.apache.hadoop.io.Writable |
AccumuloSerDe.serialize(Object o,
ObjectInspector objectInspector) |
org.apache.accumulo.core.data.Mutation |
AccumuloRowSerializer.serialize(Object obj,
ObjectInspector objInspector) |
Constructor and Description |
---|
AccumuloSerDeParameters(org.apache.hadoop.conf.Configuration conf,
Properties tableProperties,
String serdeName) |
Modifier and Type | Method and Description |
---|---|
Object |
TypedBytesSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
RegexSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
MultiDelimitSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
ObjectInspector |
TypedBytesSerDe.getObjectInspector() |
ObjectInspector |
RegexSerDe.getObjectInspector() |
ObjectInspector |
MultiDelimitSerDe.getObjectInspector() |
void |
TypedBytesSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
RegexSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
MultiDelimitSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
org.apache.hadoop.io.Writable |
TypedBytesSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
RegexSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
MultiDelimitSerDe.serialize(Object obj,
ObjectInspector objInspector) |
Modifier and Type | Method and Description |
---|---|
Object |
S3LogDeserializer.deserialize(org.apache.hadoop.io.Writable field) |
ObjectInspector |
S3LogDeserializer.getObjectInspector() |
void |
S3LogDeserializer.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
Constructor and Description |
---|
S3LogDeserializer() |
Modifier and Type | Method and Description |
---|---|
LazyObjectBase |
HBaseKeyFactory.createKey(ObjectInspector inspector)
create custom object for hbase key
|
LazyObjectBase |
DefaultHBaseKeyFactory.createKey(ObjectInspector inspector) |
T |
CompositeHBaseKeyFactory.createKey(ObjectInspector inspector) |
ObjectInspector |
HBaseKeyFactory.createKeyObjectInspector(TypeInfo type)
create custom object inspector for hbase key
|
ObjectInspector |
DefaultHBaseKeyFactory.createKeyObjectInspector(TypeInfo type) |
protected LazyObjectBase |
LazyHBaseRow.createLazyField(int fieldID,
StructField fieldRef) |
static ObjectInspector |
HBaseLazyObjectFactory.createLazyHBaseStructInspector(LazySerDeParameters serdeParams,
int index,
HBaseKeyFactory keyFactory,
List<HBaseValueFactory> valueFactories) |
Object |
HBaseSerDe.deserialize(org.apache.hadoop.io.Writable result)
Deserialize a row from the HBase Result writable to a LazyObject
|
static void |
HBaseSerDeHelper.generateColumnTypes(Properties tbl,
List<ColumnMappings.ColumnMapping> columnsMapping,
StringBuilder sb,
org.apache.hadoop.conf.Configuration conf)
Autogenerates the column types from the given serialization class
|
ObjectInspector |
HBaseSerDe.getObjectInspector() |
static org.apache.avro.Schema |
HBaseSerDeHelper.getSchemaFromFS(String schemaFSUrl,
org.apache.hadoop.conf.Configuration conf)
Read the schema from the given hdfs url for the schema
|
void |
HBaseKeyFactory.init(HBaseSerDeParameters hbaseParam,
Properties properties)
initialize factory with properties
|
void |
DefaultHBaseKeyFactory.init(HBaseSerDeParameters hbaseParam,
Properties properties) |
void |
AbstractHBaseKeyFactory.init(HBaseSerDeParameters hbaseParam,
Properties properties) |
void |
HBaseSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Initialize the SerDe given parameters.
|
static ColumnMappings |
HBaseSerDe.parseColumnsMapping(String columnsMappingSpec) |
static ColumnMappings |
HBaseSerDe.parseColumnsMapping(String columnsMappingSpec,
boolean doColumnRegexMatching)
Parses the HBase columns mapping specifier to identify the column families, qualifiers
and also caches the byte arrays corresponding to them.
|
org.apache.hadoop.io.Writable |
HBaseSerDe.serialize(Object obj,
ObjectInspector objInspector) |
Constructor and Description |
---|
HBaseSerDe() |
Modifier and Type | Method and Description |
---|---|
LazyObjectBase |
StructHBaseValueFactory.createValueObject(ObjectInspector inspector) |
LazyObjectBase |
HBaseValueFactory.createValueObject(ObjectInspector inspector)
create custom object for hbase value
|
LazyObjectBase |
DefaultHBaseValueFactory.createValueObject(ObjectInspector inspector) |
ObjectInspector |
HBaseValueFactory.createValueObjectInspector(TypeInfo type)
create custom object inspector for the value
|
ObjectInspector |
DefaultHBaseValueFactory.createValueObjectInspector(TypeInfo type) |
ObjectInspector |
AvroHBaseValueFactory.createValueObjectInspector(TypeInfo type) |
void |
HBaseValueFactory.init(HBaseSerDeParameters hbaseParam,
org.apache.hadoop.conf.Configuration conf,
Properties properties)
Initialize factory with properties
|
void |
DefaultHBaseValueFactory.init(HBaseSerDeParameters hbaseParams,
org.apache.hadoop.conf.Configuration conf,
Properties properties) |
void |
AvroHBaseValueFactory.init(HBaseSerDeParameters hbaseParams,
org.apache.hadoop.conf.Configuration conf,
Properties properties) |
Modifier and Type | Method and Description |
---|---|
static List<FieldSchema> |
MetaStoreUtils.getFieldsFromDeserializer(String tableName,
Deserializer deserializer) |
Modifier and Type | Method and Description |
---|---|
protected void |
MapJoinOperator.reloadHashTable(byte pos,
int partitionId)
Reload hashtable from the hash partition.
|
static StructObjectInspector |
PTFPartition.setupPartitionOutputOI(SerDe serDe,
StructObjectInspector tblFnOI) |
protected HiveKey |
ReduceSinkOperator.toHiveKey(Object obj,
int tag,
Integer distLength) |
Modifier and Type | Method and Description |
---|---|
int |
MapJoinBytesTableContainer.KeyValueHelper.getHashFromKey()
Get hash value from the key.
|
int |
MapJoinBytesTableContainer.DirectKeyValueWriter.getHashFromKey() |
void |
BytesBytesMultiHashMap.put(BytesBytesMultiHashMap.KvSource kv,
int keyHashCode) |
void |
MapJoinTableContainerDirectAccess.put(org.apache.hadoop.io.Writable currentKey,
org.apache.hadoop.io.Writable currentValue) |
void |
MapJoinBytesTableContainer.put(org.apache.hadoop.io.Writable currentKey,
org.apache.hadoop.io.Writable currentValue) |
void |
HybridHashTableContainer.put(org.apache.hadoop.io.Writable currentKey,
org.apache.hadoop.io.Writable currentValue) |
MapJoinKey |
MapJoinTableContainer.putRow(MapJoinObjectSerDeContext keyContext,
org.apache.hadoop.io.Writable currentKey,
MapJoinObjectSerDeContext valueContext,
org.apache.hadoop.io.Writable currentValue)
Adds row from input to the table.
|
MapJoinKey |
MapJoinBytesTableContainer.putRow(MapJoinObjectSerDeContext keyContext,
org.apache.hadoop.io.Writable currentKey,
MapJoinObjectSerDeContext valueContext,
org.apache.hadoop.io.Writable currentValue) |
MapJoinKey |
HybridHashTableContainer.putRow(MapJoinObjectSerDeContext keyContext,
org.apache.hadoop.io.Writable currentKey,
MapJoinObjectSerDeContext valueContext,
org.apache.hadoop.io.Writable currentValue) |
MapJoinKey |
HashMapWrapper.putRow(MapJoinObjectSerDeContext keyContext,
org.apache.hadoop.io.Writable currentKey,
MapJoinObjectSerDeContext valueContext,
org.apache.hadoop.io.Writable currentValue) |
static MapJoinKey |
MapJoinKey.read(ByteStream.Output output,
MapJoinObjectSerDeContext context,
org.apache.hadoop.io.Writable writable) |
void |
MapJoinKeyObject.read(MapJoinObjectSerDeContext context,
ObjectInputStream in,
org.apache.hadoop.io.Writable container) |
void |
MapJoinEagerRowContainer.read(MapJoinObjectSerDeContext context,
ObjectInputStream in,
org.apache.hadoop.io.Writable container) |
void |
MapJoinKeyObject.read(MapJoinObjectSerDeContext context,
org.apache.hadoop.io.Writable container) |
void |
MapJoinEagerRowContainer.read(MapJoinObjectSerDeContext context,
org.apache.hadoop.io.Writable currentValue) |
protected void |
MapJoinKeyObject.read(ObjectInspector oi,
Object obj) |
static ByteStream.Output |
MapJoinKey.serializeVector(ByteStream.Output byteStream,
VectorHashKeyWrapper kw,
VectorExpressionWriter[] keyOutputWriters,
VectorHashKeyWrapperBatch keyWrapperBatch,
boolean[] nulls,
boolean[] sortableSortOrders)
Serializes row to output for vectorized path.
|
void |
MapJoinBytesTableContainer.KeyValueHelper.setKeyValue(org.apache.hadoop.io.Writable key,
org.apache.hadoop.io.Writable val) |
void |
MapJoinBytesTableContainer.DirectKeyValueWriter.setKeyValue(org.apache.hadoop.io.Writable key,
org.apache.hadoop.io.Writable val) |
void |
UnwrapRowContainer.write(MapJoinObjectSerDeContext valueContext,
ObjectOutputStream out) |
void |
MapJoinRowContainer.write(MapJoinObjectSerDeContext valueContext,
ObjectOutputStream out) |
void |
MapJoinKeyObject.write(MapJoinObjectSerDeContext context,
ObjectOutputStream out) |
abstract void |
MapJoinKey.write(MapJoinObjectSerDeContext context,
ObjectOutputStream out) |
void |
MapJoinEagerRowContainer.write(MapJoinObjectSerDeContext context,
ObjectOutputStream out) |
void |
MapJoinBytesTableContainer.DirectKeyValueWriter.writeKey(ByteStream.RandomAccessOutput dest) |
void |
BytesBytesMultiHashMap.KvSource.writeKey(ByteStream.RandomAccessOutput dest)
Write key into output.
|
void |
MapJoinBytesTableContainer.DirectKeyValueWriter.writeValue(ByteStream.RandomAccessOutput dest) |
void |
BytesBytesMultiHashMap.KvSource.writeValue(ByteStream.RandomAccessOutput dest)
Write value into output.
|
Constructor and Description |
---|
HybridHashTableContainer(org.apache.hadoop.conf.Configuration hconf,
long keyCount,
long memoryAvailable,
long estimatedTableSize,
HybridHashTableConf nwayConf) |
MapJoinBytesTableContainer(org.apache.hadoop.conf.Configuration hconf,
MapJoinObjectSerDeContext valCtx,
long keyCount,
long memUsage) |
MapJoinObjectSerDeContext(SerDe serde,
boolean hasFilter) |
Modifier and Type | Method and Description |
---|---|
protected org.apache.hadoop.hive.ql.exec.tez.DynamicPartitionPruner.SourceInfo |
DynamicPartitionPruner.createSourceInfo(TableDesc t,
ExprNodeDesc partKeyExpr,
String columnName,
org.apache.hadoop.mapred.JobConf jobConf) |
protected String |
DynamicPartitionPruner.processPayload(ByteBuffer payload,
String sourceName) |
void |
DynamicPartitionPruner.prune() |
Constructor and Description |
---|
DynamicPartitionPruner(org.apache.tez.runtime.api.InputInitializerContext context,
MapWork work,
org.apache.hadoop.mapred.JobConf jobConf) |
HiveSplitGenerator(org.apache.tez.runtime.api.InputInitializerContext initializerContext) |
Modifier and Type | Method and Description |
---|---|
void |
VectorizedRowBatchCtx.addRowToBatch(int rowIndex,
org.apache.hadoop.io.Writable rowBlob,
VectorizedRowBatch batch,
org.apache.hadoop.io.DataOutputBuffer buffer)
Adds the row to the batch after deserializing the row
|
void |
VectorizedRowBatchCtx.convertRowBatchBlobToVectorizedBatch(Object rowBlob,
int rowsInBlob,
VectorizedRowBatch batch)
Deserialized set of rows and populates the batch
|
Object |
VectorizedColumnarSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
void |
VectorizedSerde.deserializeVector(Object rowBlob,
int rowsInBlob,
VectorizedRowBatch reuseBatch) |
void |
VectorizedColumnarSerDe.deserializeVector(Object rowBlob,
int rowsInBlob,
VectorizedRowBatch reuseBatch)
Deserializes the rowBlob into Vectorized row batch
|
ObjectInspector |
VectorizedColumnarSerDe.getObjectInspector() |
void |
VectorizedRowBatchCtx.init(org.apache.hadoop.conf.Configuration hiveConf,
org.apache.hadoop.mapred.FileSplit split)
Initializes VectorizedRowBatch context based on the
split and Hive configuration (Job conf with hive Plan).
|
org.apache.hadoop.io.Writable |
VectorizedColumnarSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
VectorizedSerde.serializeVector(VectorizedRowBatch vrg,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
VectorizedColumnarSerDe.serializeVector(VectorizedRowBatch vrg,
ObjectInspector objInspector)
Serialize a vectorized row batch
|
Constructor and Description |
---|
VectorizedColumnarSerDe() |
Modifier and Type | Method and Description |
---|---|
protected void |
VectorMapJoinGenerateResultOperator.reloadHashTable(byte pos,
int partitionId) |
Modifier and Type | Method and Description |
---|---|
MapJoinKey |
VectorMapJoinFastTableContainer.putRow(MapJoinObjectSerDeContext keyContext,
org.apache.hadoop.io.Writable currentKey,
MapJoinObjectSerDeContext valueContext,
org.apache.hadoop.io.Writable currentValue) |
Constructor and Description |
---|
VectorMapJoinFastTableContainer(MapJoinDesc desc,
org.apache.hadoop.conf.Configuration hconf,
long keyCount) |
Modifier and Type | Method and Description |
---|---|
void |
VectorMapJoinHashTable.putRow(org.apache.hadoop.io.BytesWritable currentKey,
org.apache.hadoop.io.BytesWritable currentValue) |
Modifier and Type | Method and Description |
---|---|
void |
VectorMapJoinOptimizedHashTable.putRow(org.apache.hadoop.io.BytesWritable currentKey,
org.apache.hadoop.io.BytesWritable currentValue) |
protected void |
VectorMapJoinOptimizedHashTable.putRowInternal(org.apache.hadoop.io.BytesWritable key,
org.apache.hadoop.io.BytesWritable value) |
Modifier and Type | Method and Description |
---|---|
Object |
OrcSerde.deserialize(org.apache.hadoop.io.Writable writable) |
void |
OrcSerde.deserializeVector(Object rowBlob,
int rowsInBatch,
VectorizedRowBatch reuseBatch) |
ObjectInspector |
OrcSerde.getObjectInspector() |
org.apache.hadoop.io.Writable |
OrcSerde.serializeVector(VectorizedRowBatch vrg,
ObjectInspector objInspector) |
Modifier and Type | Method and Description |
---|---|
Object |
ParquetHiveSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
ObjectInspector |
ParquetHiveSerDe.getObjectInspector() |
void |
ParquetHiveSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
org.apache.hadoop.io.Writable |
ParquetHiveSerDe.serialize(Object obj,
ObjectInspector objInspector) |
Modifier and Type | Method and Description |
---|---|
protected static SerDe |
PTFTranslator.createLazyBinarySerDe(org.apache.hadoop.conf.Configuration cfg,
StructObjectInspector oi,
Map<String,String> serdePropsMap) |
Modifier and Type | Method and Description |
---|---|
Object |
TypedSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
RegexSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
OpenCSVSerde.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
NullStructSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
Object |
MetadataTypedColumnsetSerDe.deserialize(org.apache.hadoop.io.Writable field) |
Object |
Deserializer.deserialize(org.apache.hadoop.io.Writable blob)
Deprecated.
Deserialize an object out of a Writable blob.
|
Object |
ByteStreamTypedSerDe.deserialize(org.apache.hadoop.io.Writable field) |
abstract Object |
AbstractSerDe.deserialize(org.apache.hadoop.io.Writable blob)
Deserialize an object out of a Writable blob.
|
Object |
AbstractEncodingAwareSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
abstract Object |
AbstractDeserializer.deserialize(org.apache.hadoop.io.Writable blob)
Deserialize an object out of a Writable blob.
|
Object |
DelimitedJSONSerDe.doDeserialize(org.apache.hadoop.io.Writable field)
Not implemented.
|
protected abstract Object |
AbstractEncodingAwareSerDe.doDeserialize(org.apache.hadoop.io.Writable blob) |
protected abstract org.apache.hadoop.io.Writable |
AbstractEncodingAwareSerDe.doSerialize(Object obj,
ObjectInspector objInspector) |
ObjectInspector |
TypedSerDe.getObjectInspector() |
ObjectInspector |
RegexSerDe.getObjectInspector() |
ObjectInspector |
OpenCSVSerde.getObjectInspector() |
ObjectInspector |
NullStructSerDe.getObjectInspector() |
ObjectInspector |
MetadataTypedColumnsetSerDe.getObjectInspector() |
ObjectInspector |
Deserializer.getObjectInspector()
Deprecated.
Get the object inspector that can be used to navigate through the internal
structure of the Object returned from deserialize(...).
|
abstract ObjectInspector |
AbstractSerDe.getObjectInspector()
Get the object inspector that can be used to navigate through the internal
structure of the Object returned from deserialize(...).
|
abstract ObjectInspector |
AbstractDeserializer.getObjectInspector()
Get the object inspector that can be used to navigate through the internal
structure of the Object returned from deserialize(...).
|
void |
TypedSerDe.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
void |
Serializer.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Deprecated.
Initialize the HiveSerializer.
|
void |
RegexSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
OpenCSVSerde.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
NullStructSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
MetadataTypedColumnsetSerDe.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
void |
Deserializer.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Deprecated.
Initialize the HiveDeserializer.
|
abstract void |
AbstractSerializer.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Initialize the HiveSerializer.
|
abstract void |
AbstractSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Deprecated.
|
void |
AbstractEncodingAwareSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Deprecated.
|
abstract void |
AbstractDeserializer.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Initialize the HiveDeserializer.
|
void |
AbstractSerDe.initialize(org.apache.hadoop.conf.Configuration configuration,
Properties tableProperties,
Properties partitionProperties)
Initialize the SerDe.
|
static void |
SerDeUtils.initializeSerDe(Deserializer deserializer,
org.apache.hadoop.conf.Configuration conf,
Properties tblProps,
Properties partProps)
Initializes a SerDe.
|
static void |
SerDeUtils.initializeSerDeWithoutErrorCheck(Deserializer deserializer,
org.apache.hadoop.conf.Configuration conf,
Properties tblProps,
Properties partProps)
Initializes a SerDe.
|
org.apache.hadoop.io.Writable |
TypedSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
Serializer.serialize(Object obj,
ObjectInspector objInspector)
Deprecated.
Serialize an object by navigating inside the Object with the
ObjectInspector.
|
org.apache.hadoop.io.Writable |
RegexSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
OpenCSVSerde.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
NullStructSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
MetadataTypedColumnsetSerDe.serialize(Object obj,
ObjectInspector objInspector) |
abstract org.apache.hadoop.io.Writable |
AbstractSerializer.serialize(Object obj,
ObjectInspector objInspector)
Serialize an object by navigating inside the Object with the
ObjectInspector.
|
abstract org.apache.hadoop.io.Writable |
AbstractSerDe.serialize(Object obj,
ObjectInspector objInspector)
Serialize an object by navigating inside the Object with the
ObjectInspector.
|
org.apache.hadoop.io.Writable |
AbstractEncodingAwareSerDe.serialize(Object obj,
ObjectInspector objInspector) |
protected void |
DelimitedJSONSerDe.serializeField(ByteStream.Output out,
Object obj,
ObjectInspector objInspector,
LazySerDeParameters serdeParams) |
Constructor and Description |
---|
ByteStreamTypedSerDe(Type objectType) |
DelimitedJSONSerDe() |
MetadataTypedColumnsetSerDe() |
TypedSerDe(Type objectType) |
Modifier and Type | Class and Description |
---|---|
class |
AvroSerdeException |
class |
BadSchemaException |
Modifier and Type | Method and Description |
---|---|
Object |
AvroSerDe.deserialize(org.apache.hadoop.io.Writable writable) |
ObjectInspector |
AvroSerDe.getObjectInspector() |
void |
AvroSerDe.initialize(org.apache.hadoop.conf.Configuration configuration,
Properties properties) |
void |
AvroSerDe.initialize(org.apache.hadoop.conf.Configuration configuration,
Properties tableProperties,
Properties partitionProperties) |
org.apache.hadoop.io.Writable |
AvroSerDe.serialize(Object o,
ObjectInspector objectInspector) |
Constructor and Description |
---|
AvroObjectInspectorGenerator(org.apache.avro.Schema schema) |
Modifier and Type | Method and Description |
---|---|
Object |
BinarySortableSerDe.deserialize(org.apache.hadoop.io.Writable blob) |
ObjectInspector |
BinarySortableSerDe.getObjectInspector() |
void |
BinarySortableSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
org.apache.hadoop.io.Writable |
BinarySortableSerDe.serialize(Object obj,
ObjectInspector objInspector) |
static void |
BinarySortableSerDe.serializeStruct(ByteStream.Output byteStream,
Object[] fieldData,
List<ObjectInspector> fieldOis,
boolean[] sortableSortOrders) |
Modifier and Type | Method and Description |
---|---|
Object |
ColumnarSerDeBase.deserialize(org.apache.hadoop.io.Writable blob) |
ObjectInspector |
ColumnarSerDeBase.getObjectInspector() |
void |
LazyBinaryColumnarSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
ColumnarSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Initialize the SerDe given the parameters.
|
protected void |
ColumnarSerDeBase.initialize(int size) |
org.apache.hadoop.io.Writable |
LazyBinaryColumnarSerDe.serialize(Object obj,
ObjectInspector objInspector) |
org.apache.hadoop.io.Writable |
ColumnarSerDe.serialize(Object obj,
ObjectInspector objInspector)
Serialize a row of data.
|
Constructor and Description |
---|
ColumnarSerDe() |
Modifier and Type | Method and Description |
---|---|
Object |
DynamicSerDeTypeString.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypeSet.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot)
NOTE: Set is not supported by Hive yet.
|
Map<Object,Object> |
DynamicSerDeTypeMap.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
ArrayList<Object> |
DynamicSerDeTypeList.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypei64.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypei32.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypei16.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypeDouble.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypedef.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypeByte.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeTypeBool.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
abstract Object |
DynamicSerDeTypeBase.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeStructBase.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDeFieldList.deserialize(Object reuse,
org.apache.thrift.protocol.TProtocol iprot) |
String |
DynamicSerDeTypeString.deserialize(org.apache.thrift.protocol.TProtocol iprot) |
Byte |
DynamicSerDeTypeByte.deserialize(org.apache.thrift.protocol.TProtocol iprot) |
Object |
DynamicSerDe.deserialize(org.apache.hadoop.io.Writable field) |
static ObjectInspector |
DynamicSerDe.dynamicSerDeStructBaseToObjectInspector(DynamicSerDeTypeBase bt) |
ObjectInspector |
DynamicSerDe.getObjectInspector() |
Class |
DynamicSerDeTypeBase.getRealType() |
void |
DynamicSerDe.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
org.apache.hadoop.io.Writable |
DynamicSerDe.serialize(Object obj,
ObjectInspector objInspector) |
void |
DynamicSerDeTypeString.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeSet.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeMap.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeList.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypei64.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypei32.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypei16.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeDouble.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypedef.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeByte.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeTypeBool.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
abstract void |
DynamicSerDeTypeBase.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
void |
DynamicSerDeStructBase.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot)
serialize
The way to serialize a Thrift "table" which in thrift land is really a
function and thus this class's name.
|
void |
DynamicSerDeFieldList.serialize(Object o,
ObjectInspector oi,
org.apache.thrift.protocol.TProtocol oprot) |
Modifier and Type | Method and Description |
---|---|
static ObjectInspector |
LazyFactory.createColumnarStructInspector(List<String> columnNames,
List<TypeInfo> columnTypes,
byte[] separators,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar)
Deprecated.
|
static ObjectInspector |
LazyFactory.createColumnarStructInspector(List<String> columnNames,
List<TypeInfo> columnTypes,
LazyObjectInspectorParameters lazyParams)
Create a hierarchical ObjectInspector for ColumnarStruct with the given
columnNames and columnTypeInfos.
|
protected LazyObjectBase |
LazyStruct.createLazyField(int fieldID,
StructField fieldRef) |
static ObjectInspector |
LazyFactory.createLazyObjectInspector(TypeInfo typeInfo,
byte[] separators,
int separatorIndex,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyObjectInspector(TypeInfo typeInfo,
byte[] separators,
int separatorIndex,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar,
boolean extendedBooleanLiteral)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyObjectInspector(TypeInfo typeInfo,
byte[] separators,
int separatorIndex,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar,
boolean extendedBooleanLiteral,
ObjectInspectorFactory.ObjectInspectorOptions option)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyObjectInspector(TypeInfo typeInfo,
byte[] separators,
int separatorIndex,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar,
ObjectInspectorFactory.ObjectInspectorOptions option)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyObjectInspector(TypeInfo typeInfo,
int separatorIndex,
LazyObjectInspectorParameters lazyParams,
ObjectInspectorFactory.ObjectInspectorOptions option)
Create a hierarchical ObjectInspector for LazyObject with the given typeInfo.
|
static ObjectInspector |
LazyFactory.createLazyStructInspector(List<String> columnNames,
List<TypeInfo> typeInfos,
byte[] separators,
org.apache.hadoop.io.Text nullSequence,
boolean lastColumnTakesRest,
boolean escaped,
byte escapeChar)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyStructInspector(List<String> columnNames,
List<TypeInfo> typeInfos,
byte[] separators,
org.apache.hadoop.io.Text nullSequence,
boolean lastColumnTakesRest,
boolean escaped,
byte escapeChar,
boolean extendedBooleanLiteral)
Deprecated.
|
static ObjectInspector |
LazyFactory.createLazyStructInspector(List<String> columnNames,
List<TypeInfo> typeInfos,
LazyObjectInspectorParameters lazyParams)
Create a hierarchical ObjectInspector for LazyStruct with the given
columnNames and columnTypeInfos.
|
Object |
LazySimpleSerDe.doDeserialize(org.apache.hadoop.io.Writable field)
Deserialize a row from the Writable to a LazyObject.
|
org.apache.hadoop.io.Writable |
LazySimpleSerDe.doSerialize(Object obj,
ObjectInspector objInspector)
Serialize a row of data.
|
void |
LazySerDeParameters.extractColumnInfo()
Extracts and set column names and column types from the table properties
|
ObjectInspector |
LazySimpleSerDe.getObjectInspector()
Returns the ObjectInspector for the row.
|
void |
LazySimpleSerDe.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl)
Initialize the SerDe given the parameters.
|
static void |
LazySimpleSerDe.serialize(ByteStream.Output out,
Object obj,
ObjectInspector objInspector,
byte[] separators,
int level,
org.apache.hadoop.io.Text nullSequence,
boolean escaped,
byte escapeChar,
boolean[] needsEscape)
Serialize the row into the StringBuilder.
|
protected void |
LazySimpleSerDe.serializeField(ByteStream.Output out,
Object obj,
ObjectInspector objInspector,
LazySerDeParameters serdeParams) |
void |
LazySerDeParameters.setSeparator(int index,
byte separator) |
Constructor and Description |
---|
LazySerDeParameters(org.apache.hadoop.conf.Configuration job,
Properties tbl,
String serdeName) |
LazySimpleSerDe() |
Modifier and Type | Method and Description |
---|---|
Object |
LazyBinarySerDe.deserialize(org.apache.hadoop.io.Writable field)
Deserialize a table record to a lazybinary struct.
|
ObjectInspector |
LazyBinarySerDe.getObjectInspector()
Returns the ObjectInspector for the row.
|
void |
LazyBinarySerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl)
Initialize the SerDe with configuration and table information.
|
static void |
LazyBinarySerDe.serialize(ByteStream.RandomAccessOutput byteStream,
Object obj,
ObjectInspector objInspector,
boolean skipLengthPrefix,
LazyBinarySerDe.BooleanRef warnedOnceNullMapKey)
A recursive function that serialize an object to a byte buffer based on its
object inspector.
|
org.apache.hadoop.io.Writable |
LazyBinarySerDe.serialize(Object obj,
ObjectInspector objInspector)
Serialize an object to a byte buffer in a binary compact way.
|
static void |
LazyBinarySerDe.serializeStruct(ByteStream.RandomAccessOutput byteStream,
Object[] fieldData,
List<ObjectInspector> fieldOis) |
Constructor and Description |
---|
LazyBinarySerDe() |
Modifier and Type | Method and Description |
---|---|
static void |
ObjectInspectorUtils.copyStructToArray(Object o,
ObjectInspector oi,
ObjectInspectorUtils.ObjectInspectorCopyOption objectInspectorOption,
Object[] dest,
int offset) |
static int |
ObjectInspectorUtils.getStructSize(ObjectInspector oi) |
Modifier and Type | Method and Description |
---|---|
Object |
ThriftDeserializer.deserialize(org.apache.hadoop.io.Writable field) |
Object |
ThriftByteStreamTypedSerDe.deserialize(org.apache.hadoop.io.Writable field) |
ObjectInspector |
ThriftDeserializer.getObjectInspector() |
void |
ThriftDeserializer.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
void |
ThriftByteStreamTypedSerDe.initialize(org.apache.hadoop.conf.Configuration job,
Properties tbl) |
Constructor and Description |
---|
ThriftByteStreamTypedSerDe(Type objectType,
org.apache.thrift.protocol.TProtocolFactory inFactory,
org.apache.thrift.protocol.TProtocolFactory outFactory) |
Modifier and Type | Method and Description |
---|---|
Object |
JsonSerDe.deserialize(org.apache.hadoop.io.Writable blob)
Takes JSON string in Text form, and has to return an object representation above
it that's readable by the corresponding object inspector.
|
Object |
HCatRecordSerDe.deserialize(org.apache.hadoop.io.Writable data)
The purpose of a deserialize method is to turn a data blob
which is a writable representation of the data into an
object that can then be parsed using the appropriate
ObjectInspector.
|
static HCatRecordObjectInspector |
HCatRecordObjectInspectorFactory.getHCatRecordObjectInspector(StructTypeInfo typeInfo)
Returns HCatRecordObjectInspector given a StructTypeInfo type definition for the record to look into
|
ObjectInspector |
JsonSerDe.getObjectInspector()
Returns an object inspector for the specified schema that
is capable of reading in the object representation of the JSON string
|
ObjectInspector |
HCatRecordSerDe.getObjectInspector()
Return an object inspector that can read through the object
that we return from deserialize().
|
void |
JsonSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
HCatRecordSerDe.initialize(org.apache.hadoop.conf.Configuration conf,
Properties tbl) |
void |
HCatRecordSerDe.initialize(HCatSchema hsch) |
org.apache.hadoop.io.Writable |
JsonSerDe.serialize(Object obj,
ObjectInspector objInspector)
Given an object and object inspector pair, traverse the object
and generate a Text representation of the object.
|
org.apache.hadoop.io.Writable |
HCatRecordSerDe.serialize(Object obj,
ObjectInspector objInspector)
The purpose of the serialize method is to turn an object-representation
with a provided ObjectInspector into a Writable format, which
the underlying layer can then use to write out.
|
static Object |
HCatRecordSerDe.serializeField(Object field,
ObjectInspector fieldObjectInspector)
Return underlying Java Object from an object-representation
that is readable by a provided ObjectInspector.
|
Constructor and Description |
---|
HCatRecordSerDe() |
Copyright © 2017 The Apache Software Foundation. All rights reserved.