public class MapredParquetOutputFormat extends org.apache.hadoop.mapred.FileOutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord> implements HiveOutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>
| Modifier and Type | Field and Description |
|---|---|
protected org.apache.parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> |
realOutputFormat |
| Constructor and Description |
|---|
MapredParquetOutputFormat() |
MapredParquetOutputFormat(org.apache.hadoop.mapreduce.OutputFormat<Void,ParquetHiveRecord> mapreduceOutputFormat) |
| Modifier and Type | Method and Description |
|---|---|
void |
checkOutputSpecs(org.apache.hadoop.fs.FileSystem ignored,
org.apache.hadoop.mapred.JobConf job) |
FileSinkOperator.RecordWriter |
getHiveRecordWriter(org.apache.hadoop.mapred.JobConf jobConf,
org.apache.hadoop.fs.Path finalOutPath,
Class<? extends org.apache.hadoop.io.Writable> valueClass,
boolean isCompressed,
Properties tableProperties,
org.apache.hadoop.util.Progressable progress)
Create the parquet schema from the hive schema, and return the RecordWriterWrapper which
contains the real output format
|
protected ParquetRecordWriterWrapper |
getParquerRecordWriterWrapper(org.apache.parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat,
org.apache.hadoop.mapred.JobConf jobConf,
String finalOutPath,
org.apache.hadoop.util.Progressable progress,
Properties tableProperties) |
org.apache.hadoop.mapred.RecordWriter<org.apache.hadoop.io.NullWritable,ParquetHiveRecord> |
getRecordWriter(org.apache.hadoop.fs.FileSystem ignored,
org.apache.hadoop.mapred.JobConf job,
String name,
org.apache.hadoop.util.Progressable progress) |
protected org.apache.parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat
public MapredParquetOutputFormat()
public MapredParquetOutputFormat(org.apache.hadoop.mapreduce.OutputFormat<Void,ParquetHiveRecord> mapreduceOutputFormat)
public void checkOutputSpecs(org.apache.hadoop.fs.FileSystem ignored,
org.apache.hadoop.mapred.JobConf job)
throws IOException
checkOutputSpecs in interface org.apache.hadoop.mapred.OutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>checkOutputSpecs in class org.apache.hadoop.mapred.FileOutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>IOExceptionpublic org.apache.hadoop.mapred.RecordWriter<org.apache.hadoop.io.NullWritable,ParquetHiveRecord> getRecordWriter(org.apache.hadoop.fs.FileSystem ignored, org.apache.hadoop.mapred.JobConf job, String name, org.apache.hadoop.util.Progressable progress) throws IOException
getRecordWriter in interface org.apache.hadoop.mapred.OutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>getRecordWriter in class org.apache.hadoop.mapred.FileOutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>ignored - Unused parameterjob - JobConf - expecting mandatory parameter PARQUET_HIVE_SCHEMAname - Path to write toprogress - ProgressIOExceptionpublic FileSinkOperator.RecordWriter getHiveRecordWriter(org.apache.hadoop.mapred.JobConf jobConf, org.apache.hadoop.fs.Path finalOutPath, Class<? extends org.apache.hadoop.io.Writable> valueClass, boolean isCompressed, Properties tableProperties, org.apache.hadoop.util.Progressable progress) throws IOException
getHiveRecordWriter in interface HiveOutputFormat<org.apache.hadoop.io.NullWritable,ParquetHiveRecord>jobConf - the job configuration filefinalOutPath - the final output file to be createdvalueClass - the value class used for createisCompressed - whether the content is compressed or nottableProperties - the table properties of this file's corresponding tableprogress - progress used for status reportIOExceptionprotected ParquetRecordWriterWrapper getParquerRecordWriterWrapper(org.apache.parquet.hadoop.ParquetOutputFormat<ParquetHiveRecord> realOutputFormat, org.apache.hadoop.mapred.JobConf jobConf, String finalOutPath, org.apache.hadoop.util.Progressable progress, Properties tableProperties) throws IOException
IOExceptionCopyright © 2019 The Apache Software Foundation. All Rights Reserved.