public class MapredParquetOutputFormat extends org.apache.hadoop.mapred.FileOutputFormat<Void,ParquetHiveRecord> implements HiveOutputFormat<Void,ParquetHiveRecord>
Modifier and Type | Field and Description |
---|---|
protected parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> |
realOutputFormat |
Constructor and Description |
---|
MapredParquetOutputFormat() |
MapredParquetOutputFormat(org.apache.hadoop.mapreduce.OutputFormat<Void,ParquetHiveRecord> mapreduceOutputFormat) |
Modifier and Type | Method and Description |
---|---|
void |
checkOutputSpecs(org.apache.hadoop.fs.FileSystem ignored,
org.apache.hadoop.mapred.JobConf job) |
FileSinkOperator.RecordWriter |
getHiveRecordWriter(org.apache.hadoop.mapred.JobConf jobConf,
org.apache.hadoop.fs.Path finalOutPath,
Class<? extends org.apache.hadoop.io.Writable> valueClass,
boolean isCompressed,
Properties tableProperties,
org.apache.hadoop.util.Progressable progress)
Create the parquet schema from the hive schema, and return the RecordWriterWrapper which
contains the real output format
|
protected ParquetRecordWriterWrapper |
getParquerRecordWriterWrapper(parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat,
org.apache.hadoop.mapred.JobConf jobConf,
String finalOutPath,
org.apache.hadoop.util.Progressable progress,
Properties tableProperties) |
org.apache.hadoop.mapred.RecordWriter<Void,ParquetHiveRecord> |
getRecordWriter(org.apache.hadoop.fs.FileSystem ignored,
org.apache.hadoop.mapred.JobConf job,
String name,
org.apache.hadoop.util.Progressable progress) |
protected parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat
public MapredParquetOutputFormat()
public MapredParquetOutputFormat(org.apache.hadoop.mapreduce.OutputFormat<Void,ParquetHiveRecord> mapreduceOutputFormat)
public void checkOutputSpecs(org.apache.hadoop.fs.FileSystem ignored, org.apache.hadoop.mapred.JobConf job) throws IOException
checkOutputSpecs
in interface org.apache.hadoop.mapred.OutputFormat<Void,ParquetHiveRecord>
checkOutputSpecs
in class org.apache.hadoop.mapred.FileOutputFormat<Void,ParquetHiveRecord>
IOException
public org.apache.hadoop.mapred.RecordWriter<Void,ParquetHiveRecord> getRecordWriter(org.apache.hadoop.fs.FileSystem ignored, org.apache.hadoop.mapred.JobConf job, String name, org.apache.hadoop.util.Progressable progress) throws IOException
getRecordWriter
in interface org.apache.hadoop.mapred.OutputFormat<Void,ParquetHiveRecord>
getRecordWriter
in class org.apache.hadoop.mapred.FileOutputFormat<Void,ParquetHiveRecord>
IOException
public FileSinkOperator.RecordWriter getHiveRecordWriter(org.apache.hadoop.mapred.JobConf jobConf, org.apache.hadoop.fs.Path finalOutPath, Class<? extends org.apache.hadoop.io.Writable> valueClass, boolean isCompressed, Properties tableProperties, org.apache.hadoop.util.Progressable progress) throws IOException
getHiveRecordWriter
in interface HiveOutputFormat<Void,ParquetHiveRecord>
jobConf
- the job configuration filefinalOutPath
- the final output file to be createdvalueClass
- the value class used for createisCompressed
- whether the content is compressed or nottableProperties
- the table properties of this file's corresponding tableprogress
- progress used for status reportIOException
protected ParquetRecordWriterWrapper getParquerRecordWriterWrapper(parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat, org.apache.hadoop.mapred.JobConf jobConf, String finalOutPath, org.apache.hadoop.util.Progressable progress, Properties tableProperties) throws IOException
IOException
Copyright © 2017 The Apache Software Foundation. All rights reserved.